var/home/core/zuul-output/0000755000175000017500000000000015110777255014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111023214015457 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006413510715111023204017671 0ustar rootrootNov 24 06:47:33 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 06:47:33 crc restorecon[4742]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:33 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:34 crc restorecon[4742]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 06:47:34 crc restorecon[4742]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 06:47:35 crc kubenswrapper[4799]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 06:47:35 crc kubenswrapper[4799]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 06:47:35 crc kubenswrapper[4799]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 06:47:35 crc kubenswrapper[4799]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 06:47:35 crc kubenswrapper[4799]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 06:47:35 crc kubenswrapper[4799]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.400563 4799 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403727 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403770 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403776 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403781 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403785 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403793 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403797 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403803 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403808 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403812 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403818 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403823 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403829 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403834 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403839 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403861 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403866 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403870 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403875 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403881 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403886 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403890 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403893 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403898 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403903 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403907 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403911 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403915 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403920 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403925 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403929 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403933 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403937 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403941 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403944 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403948 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403951 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403956 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403960 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403963 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403968 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403972 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403976 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403979 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403983 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403987 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403990 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403994 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.403998 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404001 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404005 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404010 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404014 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404017 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404021 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404024 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404027 4799 feature_gate.go:330] unrecognized feature gate: Example Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404031 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404034 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404038 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404041 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404044 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404048 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404051 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404055 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404058 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404062 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404065 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404069 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404074 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.404077 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404183 4799 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404193 4799 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404200 4799 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404206 4799 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404212 4799 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404218 4799 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404225 4799 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404230 4799 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404235 4799 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404240 4799 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404245 4799 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404249 4799 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404254 4799 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404259 4799 flags.go:64] FLAG: --cgroup-root="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404263 4799 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404274 4799 flags.go:64] FLAG: --client-ca-file="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404278 4799 flags.go:64] FLAG: --cloud-config="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404282 4799 flags.go:64] FLAG: --cloud-provider="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404286 4799 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404292 4799 flags.go:64] FLAG: --cluster-domain="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404297 4799 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404301 4799 flags.go:64] FLAG: --config-dir="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404305 4799 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404310 4799 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404316 4799 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404320 4799 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404324 4799 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404329 4799 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404333 4799 flags.go:64] FLAG: --contention-profiling="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404337 4799 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404342 4799 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404346 4799 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404350 4799 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404355 4799 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404359 4799 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404364 4799 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404368 4799 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404372 4799 flags.go:64] FLAG: --enable-server="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404376 4799 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404381 4799 flags.go:64] FLAG: --event-burst="100" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404386 4799 flags.go:64] FLAG: --event-qps="50" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404390 4799 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404394 4799 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404398 4799 flags.go:64] FLAG: --eviction-hard="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404404 4799 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404410 4799 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404414 4799 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404421 4799 flags.go:64] FLAG: --eviction-soft="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404425 4799 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404429 4799 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404433 4799 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404437 4799 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404442 4799 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404446 4799 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404450 4799 flags.go:64] FLAG: --feature-gates="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404455 4799 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404459 4799 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404464 4799 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404468 4799 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404472 4799 flags.go:64] FLAG: --healthz-port="10248" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404476 4799 flags.go:64] FLAG: --help="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404481 4799 flags.go:64] FLAG: --hostname-override="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404485 4799 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404490 4799 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404494 4799 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404498 4799 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404502 4799 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404506 4799 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404510 4799 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404514 4799 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404518 4799 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404522 4799 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404526 4799 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404530 4799 flags.go:64] FLAG: --kube-reserved="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404534 4799 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404538 4799 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404542 4799 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404548 4799 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404552 4799 flags.go:64] FLAG: --lock-file="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404558 4799 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404563 4799 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404567 4799 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404573 4799 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404577 4799 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404582 4799 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404586 4799 flags.go:64] FLAG: --logging-format="text" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404590 4799 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404594 4799 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404598 4799 flags.go:64] FLAG: --manifest-url="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404602 4799 flags.go:64] FLAG: --manifest-url-header="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404608 4799 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404612 4799 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404617 4799 flags.go:64] FLAG: --max-pods="110" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404621 4799 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404626 4799 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404630 4799 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404634 4799 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404638 4799 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404642 4799 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404646 4799 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404660 4799 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404666 4799 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404671 4799 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404676 4799 flags.go:64] FLAG: --pod-cidr="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404681 4799 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404689 4799 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404693 4799 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404697 4799 flags.go:64] FLAG: --pods-per-core="0" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404702 4799 flags.go:64] FLAG: --port="10250" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404706 4799 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404711 4799 flags.go:64] FLAG: --provider-id="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404719 4799 flags.go:64] FLAG: --qos-reserved="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404724 4799 flags.go:64] FLAG: --read-only-port="10255" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404735 4799 flags.go:64] FLAG: --register-node="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404740 4799 flags.go:64] FLAG: --register-schedulable="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404745 4799 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404764 4799 flags.go:64] FLAG: --registry-burst="10" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404768 4799 flags.go:64] FLAG: --registry-qps="5" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404772 4799 flags.go:64] FLAG: --reserved-cpus="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404776 4799 flags.go:64] FLAG: --reserved-memory="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404781 4799 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404785 4799 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404790 4799 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404794 4799 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404798 4799 flags.go:64] FLAG: --runonce="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404802 4799 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404806 4799 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404810 4799 flags.go:64] FLAG: --seccomp-default="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404815 4799 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404819 4799 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404823 4799 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404827 4799 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404832 4799 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404836 4799 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404840 4799 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404861 4799 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404865 4799 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404869 4799 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404874 4799 flags.go:64] FLAG: --system-cgroups="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404879 4799 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404886 4799 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404890 4799 flags.go:64] FLAG: --tls-cert-file="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404895 4799 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404904 4799 flags.go:64] FLAG: --tls-min-version="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404909 4799 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404913 4799 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404918 4799 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404922 4799 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404927 4799 flags.go:64] FLAG: --v="2" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404936 4799 flags.go:64] FLAG: --version="false" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404942 4799 flags.go:64] FLAG: --vmodule="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404948 4799 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.404952 4799 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405045 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405050 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405054 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405058 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405061 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405065 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405068 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405072 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405075 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405079 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405082 4799 feature_gate.go:330] unrecognized feature gate: Example Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405086 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405089 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405093 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405096 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405100 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405103 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405107 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405111 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405114 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405117 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405122 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405128 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405132 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405135 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405139 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405142 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405146 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405150 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405153 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405157 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405160 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405164 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405168 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405172 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405175 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405180 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405184 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405189 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405193 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405197 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405200 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405205 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405209 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405213 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405217 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405222 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405226 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405230 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405234 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405237 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405242 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405245 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405251 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405256 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405260 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405263 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405267 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405300 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405304 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405308 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405311 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405315 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405318 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405322 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405325 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405329 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405332 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405338 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405342 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.405348 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.406254 4799 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.419455 4799 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.419508 4799 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419651 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419662 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419668 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419673 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419677 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419682 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419687 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419691 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419696 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419724 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419729 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419734 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419738 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419745 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419752 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419757 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419762 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419766 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419771 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419775 4799 feature_gate.go:330] unrecognized feature gate: Example Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419793 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419798 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419802 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419807 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419811 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419816 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419820 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419825 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419839 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419870 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419875 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419880 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419884 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419891 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419896 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419901 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419907 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419916 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419921 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419926 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419948 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419953 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419957 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419962 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419967 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419972 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419976 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419981 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419985 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419990 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.419995 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420000 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420004 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420027 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420033 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420038 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420043 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420050 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420054 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420060 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420065 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420070 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420075 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420079 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420086 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420109 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420115 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420121 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420126 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420131 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420138 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.420147 4799 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420354 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420364 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420369 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420374 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420380 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420386 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420391 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420396 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420417 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420421 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420426 4799 feature_gate.go:330] unrecognized feature gate: Example Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420431 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420436 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420443 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420450 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420455 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420460 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420465 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420471 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420475 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420497 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420502 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420507 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420512 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420518 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420524 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420529 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420534 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420539 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420543 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420548 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420553 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420574 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420579 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420584 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420590 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420596 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420601 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420606 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420611 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420616 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420621 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420625 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420629 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420651 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420656 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420660 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420667 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420673 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420678 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420682 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420687 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420692 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420697 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420702 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420706 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420711 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420735 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420741 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420747 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420751 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420756 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420761 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420766 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420770 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420775 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420780 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420784 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420788 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420808 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.420813 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.420821 4799 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.421787 4799 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.426896 4799 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.426992 4799 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.429124 4799 server.go:997] "Starting client certificate rotation" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.429144 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.429940 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-17 17:27:09.848058439 +0000 UTC Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.430016 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1306h39m34.418045792s for next certificate rotation Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.452522 4799 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.456693 4799 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.472792 4799 log.go:25] "Validated CRI v1 runtime API" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.506188 4799 log.go:25] "Validated CRI v1 image API" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.512734 4799 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.520836 4799 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-06-42-51-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.520884 4799 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.537547 4799 manager.go:217] Machine: {Timestamp:2025-11-24 06:47:35.534538169 +0000 UTC m=+1.190520683 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8403adc0-fd03-431c-8d73-b29da3901e7b BootID:5d29851c-ec7b-44cd-80ee-83d663957897 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e9:1d:db Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e9:1d:db Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:3f:10:9a Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:9a:23:0a Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:6c:59:de Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ac:02:f3 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:43:29:7f Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ae:3f:7b:d9:7e:19 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ca:e0:e0:f8:56:64 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.537875 4799 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.538070 4799 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.538584 4799 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.539152 4799 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.539220 4799 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.539554 4799 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.539618 4799 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.540391 4799 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.540447 4799 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.542045 4799 state_mem.go:36] "Initialized new in-memory state store" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.542240 4799 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.545843 4799 kubelet.go:418] "Attempting to sync node with API server" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.545904 4799 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.545945 4799 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.545977 4799 kubelet.go:324] "Adding apiserver pod source" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.545995 4799 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.550927 4799 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.553161 4799 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.556047 4799 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.558018 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.558059 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.558180 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.558180 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558610 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558646 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558654 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558662 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558675 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558682 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558690 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558701 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558711 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558719 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558731 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.558739 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.559390 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.559872 4799 server.go:1280] "Started kubelet" Nov 24 06:47:35 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.561714 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.562115 4799 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.562119 4799 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.562587 4799 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.563530 4799 server.go:460] "Adding debug handlers to kubelet server" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.565095 4799 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.565132 4799 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.565202 4799 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 04:26:11.597289135 +0000 UTC Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.565259 4799 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 333h38m36.032034315s for next certificate rotation Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.565494 4799 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.565508 4799 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.565739 4799 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.566418 4799 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.566410 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.566496 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.569571 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="200ms" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.570083 4799 factory.go:55] Registering systemd factory Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.570125 4799 factory.go:221] Registration of the systemd container factory successfully Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.571390 4799 factory.go:153] Registering CRI-O factory Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.571442 4799 factory.go:221] Registration of the crio container factory successfully Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.572042 4799 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.572078 4799 factory.go:103] Registering Raw factory Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.572094 4799 manager.go:1196] Started watching for new ooms in manager Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.570539 4799 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187ade825d3495ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 06:47:35.559812589 +0000 UTC m=+1.215795063,LastTimestamp:2025-11-24 06:47:35.559812589 +0000 UTC m=+1.215795063,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.574178 4799 manager.go:319] Starting recovery of all containers Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.581384 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.581435 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.581448 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.581460 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.581473 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.581487 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.581499 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.581509 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.581522 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.583835 4799 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.583884 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.583904 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.583917 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.583936 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.583952 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.583965 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.583978 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.583993 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584008 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584084 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584101 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584115 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584127 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584142 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584155 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584168 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584180 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584243 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584263 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584275 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584311 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584324 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584337 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584385 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584403 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584415 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584429 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584443 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584456 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584470 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584483 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584528 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584541 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584566 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584581 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584601 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584615 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584630 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584643 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584657 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584670 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584717 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584732 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584753 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584769 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584787 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584803 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584818 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584832 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584863 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584876 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584891 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584904 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584917 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584936 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584949 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584964 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584978 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.584992 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585008 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585023 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585038 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585051 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585100 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585117 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585133 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585150 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585171 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585186 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585200 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585213 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585226 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585238 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585252 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585265 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585278 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585291 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585305 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585320 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585334 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585347 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585362 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585375 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585389 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585403 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585417 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585430 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585444 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585457 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585471 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585486 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585499 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585513 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585528 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585542 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585563 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585578 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585596 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585610 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585627 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585641 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585656 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585671 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585684 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585699 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585713 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585727 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585742 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585756 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585771 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585786 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585799 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585814 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585829 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585859 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585875 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585889 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585903 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585920 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585932 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585948 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585962 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585978 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.585993 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586007 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586020 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586034 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586052 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586065 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586081 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586096 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586115 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586129 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586142 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586165 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586182 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586198 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586211 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586224 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586239 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586254 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586268 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586282 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586296 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586310 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586327 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586342 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586358 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586371 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586385 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586399 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586415 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586428 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586442 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586456 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586469 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586481 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586495 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586511 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586526 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586543 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586563 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586577 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586594 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586609 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586622 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586638 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586654 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586668 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586683 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586698 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586713 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586727 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586744 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586757 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586772 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586787 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586804 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586819 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586832 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586863 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586877 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586891 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586905 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586919 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586935 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586950 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586964 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586978 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.586992 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.587015 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.587033 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.587048 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.587108 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.587124 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.587173 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.587186 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.587202 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.587241 4799 reconstruct.go:97] "Volume reconstruction finished" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.587254 4799 reconciler.go:26] "Reconciler: start to sync state" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.594838 4799 manager.go:324] Recovery completed Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.604213 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.606181 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.606214 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.606225 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.610964 4799 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.610995 4799 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.611017 4799 state_mem.go:36] "Initialized new in-memory state store" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.623833 4799 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.626594 4799 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.626654 4799 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.626973 4799 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.627307 4799 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.628767 4799 policy_none.go:49] "None policy: Start" Nov 24 06:47:35 crc kubenswrapper[4799]: W1124 06:47:35.629666 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.629815 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.630026 4799 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.630079 4799 state_mem.go:35] "Initializing new in-memory state store" Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.666869 4799 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.699154 4799 manager.go:334] "Starting Device Plugin manager" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.699251 4799 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.699381 4799 server.go:79] "Starting device plugin registration server" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.700706 4799 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.700733 4799 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.701497 4799 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.701649 4799 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.701662 4799 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.713904 4799 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.728169 4799 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.728336 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.729871 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.729922 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.729936 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.730134 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.730590 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.730691 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.731327 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.731370 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.731387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.731527 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.731893 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.731924 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.731936 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.732259 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.732292 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.732645 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.732669 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.732679 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.732763 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.732960 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.732993 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.733343 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.733378 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.733391 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.733638 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.733661 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.733672 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.733790 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.734062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.734116 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.734137 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.734313 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.734432 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.734677 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.734703 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.734714 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.734904 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.734940 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.736395 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.736428 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.736431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.736440 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.736463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.736589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.770312 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="400ms" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790192 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790233 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790253 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790271 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790290 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790312 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790415 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790465 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790495 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790511 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790533 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790552 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790601 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790675 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.790739 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.801342 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.802940 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.802991 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.803015 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.803047 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:47:35 crc kubenswrapper[4799]: E1124 06:47:35.803736 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.891640 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.891819 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.891897 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.891904 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.891989 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892089 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892150 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892133 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892168 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892263 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892307 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892378 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892400 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892412 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892430 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892476 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892508 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892533 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892561 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892570 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892612 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892619 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892672 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892683 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892700 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892708 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892738 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892743 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892773 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:35 crc kubenswrapper[4799]: I1124 06:47:35.892893 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.004652 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.006899 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.006942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.006953 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.006977 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:47:36 crc kubenswrapper[4799]: E1124 06:47:36.007573 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.056295 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.077691 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.093738 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.101087 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.106534 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 06:47:36 crc kubenswrapper[4799]: W1124 06:47:36.107309 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-b21f32ded663e853599c3c7592a8d9db244b63247b68ed1330947a89d5c23622 WatchSource:0}: Error finding container b21f32ded663e853599c3c7592a8d9db244b63247b68ed1330947a89d5c23622: Status 404 returned error can't find the container with id b21f32ded663e853599c3c7592a8d9db244b63247b68ed1330947a89d5c23622 Nov 24 06:47:36 crc kubenswrapper[4799]: W1124 06:47:36.134417 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-43a57f3a77e972b39c6869139c5041726ff9d5c8312b97d9533d6b9e5e06b031 WatchSource:0}: Error finding container 43a57f3a77e972b39c6869139c5041726ff9d5c8312b97d9533d6b9e5e06b031: Status 404 returned error can't find the container with id 43a57f3a77e972b39c6869139c5041726ff9d5c8312b97d9533d6b9e5e06b031 Nov 24 06:47:36 crc kubenswrapper[4799]: W1124 06:47:36.139024 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-81dc7cd32651a7243b7fb1c691e1f138f1f1213afd88939c209823c6ba56a2ee WatchSource:0}: Error finding container 81dc7cd32651a7243b7fb1c691e1f138f1f1213afd88939c209823c6ba56a2ee: Status 404 returned error can't find the container with id 81dc7cd32651a7243b7fb1c691e1f138f1f1213afd88939c209823c6ba56a2ee Nov 24 06:47:36 crc kubenswrapper[4799]: W1124 06:47:36.146636 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-8b1bf80e62affe63e90e8aa447ad6478dafcb4436cc42de15d59b7da2de16b75 WatchSource:0}: Error finding container 8b1bf80e62affe63e90e8aa447ad6478dafcb4436cc42de15d59b7da2de16b75: Status 404 returned error can't find the container with id 8b1bf80e62affe63e90e8aa447ad6478dafcb4436cc42de15d59b7da2de16b75 Nov 24 06:47:36 crc kubenswrapper[4799]: W1124 06:47:36.149469 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-9ee75bd754c66685b5ef6d17d56abb75e1f9ca95a636092a8fb17d58bc0d6d26 WatchSource:0}: Error finding container 9ee75bd754c66685b5ef6d17d56abb75e1f9ca95a636092a8fb17d58bc0d6d26: Status 404 returned error can't find the container with id 9ee75bd754c66685b5ef6d17d56abb75e1f9ca95a636092a8fb17d58bc0d6d26 Nov 24 06:47:36 crc kubenswrapper[4799]: E1124 06:47:36.171792 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="800ms" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.408170 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.410740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.410796 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.410810 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.410868 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:47:36 crc kubenswrapper[4799]: E1124 06:47:36.411510 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Nov 24 06:47:36 crc kubenswrapper[4799]: E1124 06:47:36.432113 4799 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187ade825d3495ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 06:47:35.559812589 +0000 UTC m=+1.215795063,LastTimestamp:2025-11-24 06:47:35.559812589 +0000 UTC m=+1.215795063,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 06:47:36 crc kubenswrapper[4799]: W1124 06:47:36.524566 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:36 crc kubenswrapper[4799]: E1124 06:47:36.524731 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.563005 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:36 crc kubenswrapper[4799]: W1124 06:47:36.566946 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:36 crc kubenswrapper[4799]: E1124 06:47:36.567059 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.632883 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"43a57f3a77e972b39c6869139c5041726ff9d5c8312b97d9533d6b9e5e06b031"} Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.634263 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b21f32ded663e853599c3c7592a8d9db244b63247b68ed1330947a89d5c23622"} Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.635662 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9ee75bd754c66685b5ef6d17d56abb75e1f9ca95a636092a8fb17d58bc0d6d26"} Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.636690 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8b1bf80e62affe63e90e8aa447ad6478dafcb4436cc42de15d59b7da2de16b75"} Nov 24 06:47:36 crc kubenswrapper[4799]: I1124 06:47:36.637825 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"81dc7cd32651a7243b7fb1c691e1f138f1f1213afd88939c209823c6ba56a2ee"} Nov 24 06:47:36 crc kubenswrapper[4799]: W1124 06:47:36.865329 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:36 crc kubenswrapper[4799]: E1124 06:47:36.865429 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:47:36 crc kubenswrapper[4799]: W1124 06:47:36.887415 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:36 crc kubenswrapper[4799]: E1124 06:47:36.887575 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:47:36 crc kubenswrapper[4799]: E1124 06:47:36.972835 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="1.6s" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.211975 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.214556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.214633 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.214646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.214715 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:47:37 crc kubenswrapper[4799]: E1124 06:47:37.215516 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.562661 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.643605 4799 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508" exitCode=0 Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.643701 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.643686 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508"} Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.644961 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.645015 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.645033 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.646368 4799 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d" exitCode=0 Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.646454 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.646458 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d"} Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.648389 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.648418 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.648434 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.650587 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c"} Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.650641 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb"} Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.650659 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae"} Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.650673 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c"} Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.650753 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.652147 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.652550 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.652582 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.657578 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc" exitCode=0 Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.657715 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc"} Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.658061 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.664286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.664512 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f"} Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.664643 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.664786 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.664936 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.664483 4799 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f" exitCode=0 Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.666807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.666974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.667126 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.672047 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.673658 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.673682 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:37 crc kubenswrapper[4799]: I1124 06:47:37.673692 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:38 crc kubenswrapper[4799]: W1124 06:47:38.360455 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:38 crc kubenswrapper[4799]: E1124 06:47:38.361017 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:47:38 crc kubenswrapper[4799]: W1124 06:47:38.473669 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:38 crc kubenswrapper[4799]: E1124 06:47:38.473749 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.563332 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:38 crc kubenswrapper[4799]: E1124 06:47:38.574343 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="3.2s" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.668901 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6d4ddf8a6fc1ed958f9aa914d2ee1abd73dfe6c36513f0595068a4eea89d6e4c"} Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.669230 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.670141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.670172 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.670183 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.671956 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e69e47ea9ea59fe65c19fad4217224390526bb50c23984de97db2cefa6cfe0a5"} Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.671984 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b888a1477657d057e486b872a2a3da23225070bfee8b34139b33f4762c4d32fa"} Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.671998 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7b3642617df8fcafa59d2eb42352cfe1f5a455e0a702843c2fffc8491cde5c55"} Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.672073 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.672782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.672800 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.672810 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.676268 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3"} Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.676304 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721"} Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.676321 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf"} Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.677775 4799 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921" exitCode=0 Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.677912 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.678474 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.678971 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921"} Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.679579 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.679615 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.679636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.683292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.683357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.683381 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.816046 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.817635 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.817686 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.817702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:38 crc kubenswrapper[4799]: I1124 06:47:38.817735 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:47:38 crc kubenswrapper[4799]: E1124 06:47:38.818437 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Nov 24 06:47:38 crc kubenswrapper[4799]: W1124 06:47:38.827545 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Nov 24 06:47:38 crc kubenswrapper[4799]: E1124 06:47:38.827644 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.641292 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.683616 4799 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c" exitCode=0 Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.683685 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c"} Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.683779 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.684960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.685108 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.685220 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.689064 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715"} Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.689149 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050"} Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.689099 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.689262 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.689518 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.690952 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.690984 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.690998 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.691248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.691266 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.691278 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.691643 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.691656 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.691664 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.746121 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:39 crc kubenswrapper[4799]: I1124 06:47:39.983956 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.698560 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1"} Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.698634 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.698678 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329"} Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.698710 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b"} Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.698731 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c"} Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.698803 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.700423 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.700479 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.700501 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.701197 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.701259 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:40 crc kubenswrapper[4799]: I1124 06:47:40.701284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.349627 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.349986 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.352668 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.352749 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.352771 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.708798 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22"} Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.708944 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.708986 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.710688 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.710772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.710802 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.710961 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.711006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:41 crc kubenswrapper[4799]: I1124 06:47:41.711026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.019398 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.021431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.021551 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.021581 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.021634 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.511164 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.511449 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.515590 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.515660 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.515680 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.532443 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.711694 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.711707 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.713301 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.713367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.713410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.713499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.713542 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:42 crc kubenswrapper[4799]: I1124 06:47:42.713567 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.181169 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.693104 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.714654 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.714742 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.714801 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.716555 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.716611 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.716632 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.716987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.717060 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.717081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.861708 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.862044 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.863919 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.863989 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:43 crc kubenswrapper[4799]: I1124 06:47:43.864009 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:44 crc kubenswrapper[4799]: I1124 06:47:44.688464 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:44 crc kubenswrapper[4799]: I1124 06:47:44.718095 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:44 crc kubenswrapper[4799]: I1124 06:47:44.719814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:44 crc kubenswrapper[4799]: I1124 06:47:44.719921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:44 crc kubenswrapper[4799]: I1124 06:47:44.719945 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:45 crc kubenswrapper[4799]: E1124 06:47:45.714116 4799 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 06:47:46 crc kubenswrapper[4799]: I1124 06:47:46.181363 4799 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 06:47:46 crc kubenswrapper[4799]: I1124 06:47:46.181505 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 06:47:47 crc kubenswrapper[4799]: I1124 06:47:47.997188 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 06:47:47 crc kubenswrapper[4799]: I1124 06:47:47.997389 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:47 crc kubenswrapper[4799]: I1124 06:47:47.998504 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:47 crc kubenswrapper[4799]: I1124 06:47:47.998543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:47 crc kubenswrapper[4799]: I1124 06:47:47.998556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:49 crc kubenswrapper[4799]: I1124 06:47:49.563667 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 24 06:47:49 crc kubenswrapper[4799]: I1124 06:47:49.746507 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 06:47:49 crc kubenswrapper[4799]: I1124 06:47:49.746596 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 06:47:50 crc kubenswrapper[4799]: W1124 06:47:50.012249 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 24 06:47:50 crc kubenswrapper[4799]: I1124 06:47:50.012352 4799 trace.go:236] Trace[1501805467]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 06:47:40.010) (total time: 10002ms): Nov 24 06:47:50 crc kubenswrapper[4799]: Trace[1501805467]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (06:47:50.012) Nov 24 06:47:50 crc kubenswrapper[4799]: Trace[1501805467]: [10.00216437s] [10.00216437s] END Nov 24 06:47:50 crc kubenswrapper[4799]: E1124 06:47:50.012374 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 24 06:47:50 crc kubenswrapper[4799]: I1124 06:47:50.625013 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 06:47:50 crc kubenswrapper[4799]: I1124 06:47:50.625104 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 06:47:54 crc kubenswrapper[4799]: I1124 06:47:54.692604 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:54 crc kubenswrapper[4799]: I1124 06:47:54.692764 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:54 crc kubenswrapper[4799]: I1124 06:47:54.693949 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:54 crc kubenswrapper[4799]: I1124 06:47:54.694010 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:54 crc kubenswrapper[4799]: I1124 06:47:54.694023 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:54 crc kubenswrapper[4799]: I1124 06:47:54.751938 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:54 crc kubenswrapper[4799]: I1124 06:47:54.752252 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:47:54 crc kubenswrapper[4799]: I1124 06:47:54.753672 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:47:54 crc kubenswrapper[4799]: I1124 06:47:54.753722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:47:54 crc kubenswrapper[4799]: I1124 06:47:54.753739 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:47:54 crc kubenswrapper[4799]: I1124 06:47:54.756726 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.086928 4799 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.558074 4799 apiserver.go:52] "Watching apiserver" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.584161 4799 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.584702 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.585385 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.585485 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.585601 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.585628 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.585901 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.586461 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.586504 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.586561 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.586603 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.590673 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.590675 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.591048 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.591225 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.591240 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.591452 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.591470 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.591983 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.598208 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.619658 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.622015 4799 trace.go:236] Trace[1551274376]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 06:47:43.630) (total time: 11991ms): Nov 24 06:47:55 crc kubenswrapper[4799]: Trace[1551274376]: ---"Objects listed" error: 11991ms (06:47:55.621) Nov 24 06:47:55 crc kubenswrapper[4799]: Trace[1551274376]: [11.991494764s] [11.991494764s] END Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.622059 4799 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.622088 4799 trace.go:236] Trace[897583109]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 06:47:42.630) (total time: 12991ms): Nov 24 06:47:55 crc kubenswrapper[4799]: Trace[897583109]: ---"Objects listed" error: 12991ms (06:47:55.621) Nov 24 06:47:55 crc kubenswrapper[4799]: Trace[897583109]: [12.991050928s] [12.991050928s] END Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.622140 4799 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.622945 4799 trace.go:236] Trace[684881399]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 06:47:45.172) (total time: 10450ms): Nov 24 06:47:55 crc kubenswrapper[4799]: Trace[684881399]: ---"Objects listed" error: 10450ms (06:47:55.622) Nov 24 06:47:55 crc kubenswrapper[4799]: Trace[684881399]: [10.450583485s] [10.450583485s] END Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.622965 4799 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.623371 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.625266 4799 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.646034 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.661234 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.667097 4799 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.670433 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:56798->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.670456 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:56790->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.670513 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:56798->192.168.126.11:17697: read: connection reset by peer" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.670557 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:56790->192.168.126.11:17697: read: connection reset by peer" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.671189 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.671274 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.678734 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.692975 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.705676 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.725362 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726123 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726196 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726239 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726282 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726321 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726403 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726444 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726483 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726527 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726566 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726671 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726716 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726750 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726785 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726825 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.726889 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.727139 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.727213 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.727374 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.727431 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.727480 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.728247 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.728466 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730118 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730209 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730278 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730326 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730398 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730449 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730493 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730543 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730600 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730654 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730922 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731107 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731167 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731223 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731276 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731376 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731437 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731496 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731539 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731591 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731643 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731689 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731883 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731982 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732039 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732083 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732132 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732193 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732236 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732286 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732334 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732375 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732424 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732490 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732539 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732588 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.728459 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732638 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.729552 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732655 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732690 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732736 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732654 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.729834 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732790 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.729988 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732875 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730003 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.729980 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730114 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733071 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730472 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730580 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730690 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730878 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731109 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731515 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731520 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.731540 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732137 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732559 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732586 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733339 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.729810 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.730561 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733376 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733460 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.732948 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733562 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733609 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733643 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733679 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733814 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733868 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733906 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733940 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733972 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.733962 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734000 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734036 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734067 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734069 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734095 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734126 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734182 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734244 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734429 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734481 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734532 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734581 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734558 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734630 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734636 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734676 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734680 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734732 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734789 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734832 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734910 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734956 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.734959 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735032 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735072 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735160 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735200 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735234 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735278 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735311 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735345 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735366 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735388 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735424 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735459 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735486 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.735665 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.736100 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.736120 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.740942 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.740970 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.741196 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.741086 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.741286 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.741481 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.741496 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.741626 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.742403 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.742770 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.743067 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.736322 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.736481 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.736799 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.737039 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.737674 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.737739 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.738227 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.738598 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.738626 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.738635 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.738713 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.738796 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.739307 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.739445 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.739553 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.739769 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.740193 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.740231 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.740760 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.740287 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.740899 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.744272 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.744958 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.745045 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.745108 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.745134 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.745730 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.746917 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.747173 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.745916 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.745941 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.747320 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.747339 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.747443 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.747496 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.747646 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.748172 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.748350 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.748408 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.748468 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.748512 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.748525 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.748556 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.748607 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.748771 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.748820 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.749213 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.749383 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.749395 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.749561 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.749624 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.750337 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.750589 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.749837 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.751768 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.751905 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.751962 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.751996 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752217 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752242 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752265 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752288 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752234 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752312 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752337 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752363 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752391 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752392 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752415 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752443 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752472 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752496 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752523 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752546 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752573 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752602 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752625 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752649 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752671 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752697 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752726 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752757 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752785 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752814 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752871 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752895 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752922 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752946 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752971 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752998 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753025 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753052 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753079 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753103 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753129 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753154 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753179 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753207 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753231 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753258 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753287 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753310 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753378 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753411 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753437 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753464 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753491 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753597 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753626 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753652 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753683 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753720 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753760 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753792 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753823 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753875 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753903 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753927 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753953 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753980 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754004 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754033 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754058 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754127 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754151 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754177 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754207 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754251 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754277 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754304 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754329 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754363 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754393 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754425 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754456 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754508 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754537 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.755202 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.755792 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.755900 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.755932 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.755966 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.755995 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756040 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756067 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756093 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756118 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756182 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756257 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756559 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756599 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756636 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756672 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756700 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756729 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756762 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752695 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756794 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756918 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756940 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756961 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756999 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757027 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757215 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757242 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757289 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757304 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757321 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757338 4799 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757355 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757370 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757386 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757401 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757414 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757426 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757439 4799 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757451 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757462 4799 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757478 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757491 4799 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757503 4799 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757514 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757524 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757535 4799 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757547 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757559 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757572 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757583 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757594 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757605 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757616 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757631 4799 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757642 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757655 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757668 4799 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757679 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757690 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757702 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757712 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757723 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757737 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757750 4799 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757761 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757775 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757787 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757797 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757808 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757841 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757871 4799 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757884 4799 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757897 4799 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757911 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757922 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757935 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757949 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757963 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757984 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.757998 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.758012 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.758025 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.758037 4799 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.755736 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.758124 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:47:56.258074185 +0000 UTC m=+21.914056659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.758125 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.758686 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759038 4799 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759062 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759076 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759092 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759105 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759117 4799 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759128 4799 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759139 4799 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759150 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759161 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759171 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759182 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759194 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759238 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759250 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759260 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759271 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759282 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759295 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759306 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759317 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759330 4799 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759341 4799 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759351 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759362 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759374 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759387 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759586 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759597 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759607 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759619 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759629 4799 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759639 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759649 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759660 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759673 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759684 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759694 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759704 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759713 4799 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759723 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759732 4799 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759742 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759751 4799 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759762 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.759870 4799 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.760821 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.760914 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.761057 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.761198 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.761595 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.762172 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.762231 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.762411 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.762502 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.762639 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.762671 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753185 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753205 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753441 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753684 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.767173 4799 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.769794 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.768107 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715" exitCode=255 Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.763521 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.768143 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715"} Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.766399 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.769995 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.753841 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754193 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754256 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.754418 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.755403 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.755561 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756077 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756382 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.756583 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.762703 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.762732 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.763069 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.763330 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.765307 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.765878 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.765909 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.765925 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.766074 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.766243 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.766398 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.752741 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.766666 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.767134 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.770303 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:56.270270075 +0000 UTC m=+21.926252609 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.770394 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:56.270361477 +0000 UTC m=+21.926343971 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.771917 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.772300 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.773114 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.773576 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.773782 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.774082 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.790266 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.791225 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.791305 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.791996 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.792216 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.792245 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.792293 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.792317 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.792444 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:56.292384841 +0000 UTC m=+21.948367515 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.792710 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.793053 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.793290 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.793780 4799 scope.go:117] "RemoveContainer" containerID="3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.793927 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.795079 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.795911 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.796281 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.796304 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.796318 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:55 crc kubenswrapper[4799]: E1124 06:47:55.796451 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:56.296431604 +0000 UTC m=+21.952414078 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.799079 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.799106 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.799685 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.799755 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.800029 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.800074 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.800670 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.801649 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.803187 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.804138 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.804314 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.805439 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.805508 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.805757 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.805798 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.805872 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.805941 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.809061 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.809087 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.809133 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.811424 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.816239 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.816666 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.816695 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.816835 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.816896 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.817439 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.817492 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.817965 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.818109 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.818164 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.825199 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.830265 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.833335 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.833568 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.836505 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.836947 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.845590 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.846883 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.853396 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865003 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865090 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865166 4799 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865184 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865198 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865215 4799 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865233 4799 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865249 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865266 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865279 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865269 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865293 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865345 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865357 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865368 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865368 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865378 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865431 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865450 4799 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865466 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865480 4799 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865493 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865506 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865521 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865534 4799 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865548 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865561 4799 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865576 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865588 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865601 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865615 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865627 4799 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865640 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865652 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865663 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865675 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865688 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865701 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865717 4799 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865728 4799 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865740 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865752 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865764 4799 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865780 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865792 4799 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865806 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865819 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865831 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865867 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865881 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865894 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865907 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865922 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865937 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865951 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865966 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865978 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.865991 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866003 4799 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866015 4799 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866028 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866041 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866055 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866069 4799 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866082 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866094 4799 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866109 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866121 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866133 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866145 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866157 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866185 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866197 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866209 4799 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866220 4799 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866233 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866245 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866257 4799 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866276 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866289 4799 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866301 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866313 4799 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866324 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866337 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866349 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866360 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866374 4799 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866387 4799 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866398 4799 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866410 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866421 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.866433 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.880954 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.884084 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.892271 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.914955 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.919393 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.929715 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.932747 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 06:47:55 crc kubenswrapper[4799]: W1124 06:47:55.935914 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-d02fb4310d9ddd44c44d989080ee5e694f1cef66ef80e0bf7ba1b2a01778dc5c WatchSource:0}: Error finding container d02fb4310d9ddd44c44d989080ee5e694f1cef66ef80e0bf7ba1b2a01778dc5c: Status 404 returned error can't find the container with id d02fb4310d9ddd44c44d989080ee5e694f1cef66ef80e0bf7ba1b2a01778dc5c Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.941021 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: W1124 06:47:55.947818 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-4c07dce2418cea8a6bc3f52379e81f2b4d9e2a1dc76a204b42da2d626e1f0ff3 WatchSource:0}: Error finding container 4c07dce2418cea8a6bc3f52379e81f2b4d9e2a1dc76a204b42da2d626e1f0ff3: Status 404 returned error can't find the container with id 4c07dce2418cea8a6bc3f52379e81f2b4d9e2a1dc76a204b42da2d626e1f0ff3 Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.947855 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.958250 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.967443 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.967463 4799 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.973355 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:55 crc kubenswrapper[4799]: I1124 06:47:55.988728 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.005252 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.020294 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.033805 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.043924 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.179928 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.188524 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.194712 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.199460 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.215404 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.229961 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.241094 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.251689 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.261903 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.269457 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.269759 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:47:57.269706192 +0000 UTC m=+22.925688776 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.272803 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.295405 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.315322 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.331128 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.342378 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.353532 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.363638 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.370098 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.370157 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.370190 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.370221 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370330 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370370 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370386 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370393 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370417 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370350 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370443 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370463 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370443 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:57.370424831 +0000 UTC m=+23.026407305 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370529 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:57.370507843 +0000 UTC m=+23.026490487 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370553 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:57.370544884 +0000 UTC m=+23.026527588 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:47:56 crc kubenswrapper[4799]: E1124 06:47:56.370570 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:57.370563295 +0000 UTC m=+23.026546009 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.378407 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.389840 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.774807 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.776996 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e"} Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.777523 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.778101 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"54bd359f4efa08d6f2b64f0045a0c028f0fa0397d2bccbf56b6b241c7127b016"} Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.780085 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2"} Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.780123 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee"} Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.780138 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4c07dce2418cea8a6bc3f52379e81f2b4d9e2a1dc76a204b42da2d626e1f0ff3"} Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.782806 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63"} Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.782835 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d02fb4310d9ddd44c44d989080ee5e694f1cef66ef80e0bf7ba1b2a01778dc5c"} Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.790929 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.802253 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.813565 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.826481 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.840062 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.849226 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-bkb8p"] Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.849696 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-zzvdv"] Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.849931 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bkb8p" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.850549 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.850925 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-zvd7f"] Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.851390 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.856062 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.856259 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.856272 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.856422 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.857046 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-rzk7n"] Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.857393 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b7nd7"] Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.857563 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.858282 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.859610 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.861390 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.861772 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.868199 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.868681 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.869049 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.869103 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.869270 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.869363 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.869510 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.869583 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.869678 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.869769 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.869898 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.869957 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.870125 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.870276 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.870442 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.878129 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.878208 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-var-lib-cni-multus\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.878611 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-run-multus-certs\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.878635 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz6b6\" (UniqueName: \"kubernetes.io/projected/7980021c-dde3-4c14-a7b6-fbcc947a183d-kube-api-access-jz6b6\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.878666 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-env-overrides\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.878690 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-os-release\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.878711 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-conf-dir\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.878899 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-system-cni-dir\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.878948 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-var-lib-openvswitch\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.878987 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-config\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879023 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-socket-dir-parent\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879055 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-hostroot\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879106 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e2ca5e21-6366-493a-b14b-b71baa6b54d7-hosts-file\") pod \"node-resolver-bkb8p\" (UID: \"e2ca5e21-6366-493a-b14b-b71baa6b54d7\") " pod="openshift-dns/node-resolver-bkb8p" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879233 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-cnibin\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879277 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879313 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-netns\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879355 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-openvswitch\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879403 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-ovn-kubernetes\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879425 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-script-lib\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879447 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-etc-kubernetes\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879581 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-os-release\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879610 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/354ebd50-7a67-481d-86db-4b3bf4753161-rootfs\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879630 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-systemd\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879652 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-log-socket\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879673 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-var-lib-cni-bin\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879698 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-daemon-config\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879720 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879778 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-system-cni-dir\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879839 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-run-netns\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879895 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-var-lib-kubelet\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.879963 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ed80eb00-6567-4878-91c2-979bb8e2f75c-cni-binary-copy\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880169 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/354ebd50-7a67-481d-86db-4b3bf4753161-proxy-tls\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880204 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-node-log\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880231 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-netd\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880258 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-cnibin\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880299 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-kubelet\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880327 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-ovn\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880355 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovn-node-metrics-cert\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880385 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq4cv\" (UniqueName: \"kubernetes.io/projected/e2ca5e21-6366-493a-b14b-b71baa6b54d7-kube-api-access-pq4cv\") pod \"node-resolver-bkb8p\" (UID: \"e2ca5e21-6366-493a-b14b-b71baa6b54d7\") " pod="openshift-dns/node-resolver-bkb8p" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880415 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8djp\" (UniqueName: \"kubernetes.io/projected/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-kube-api-access-r8djp\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880443 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-run-k8s-cni-cncf-io\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5ttq\" (UniqueName: \"kubernetes.io/projected/ed80eb00-6567-4878-91c2-979bb8e2f75c-kube-api-access-s5ttq\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880548 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7980021c-dde3-4c14-a7b6-fbcc947a183d-cni-binary-copy\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880575 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ed80eb00-6567-4878-91c2-979bb8e2f75c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880603 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-slash\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880634 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-etc-openvswitch\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880662 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-bin\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880688 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-cni-dir\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880742 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-systemd-units\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880784 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/354ebd50-7a67-481d-86db-4b3bf4753161-mcd-auth-proxy-config\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.880817 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxr2h\" (UniqueName: \"kubernetes.io/projected/354ebd50-7a67-481d-86db-4b3bf4753161-kube-api-access-gxr2h\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.892704 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.905460 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.917428 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.929347 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.941396 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.951883 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.963184 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.973592 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981407 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-system-cni-dir\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981464 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-var-lib-openvswitch\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981489 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-config\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981516 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-socket-dir-parent\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981542 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-hostroot\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981577 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e2ca5e21-6366-493a-b14b-b71baa6b54d7-hosts-file\") pod \"node-resolver-bkb8p\" (UID: \"e2ca5e21-6366-493a-b14b-b71baa6b54d7\") " pod="openshift-dns/node-resolver-bkb8p" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981597 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-var-lib-openvswitch\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981607 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-cnibin\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981649 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-cnibin\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981624 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-system-cni-dir\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981709 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e2ca5e21-6366-493a-b14b-b71baa6b54d7-hosts-file\") pod \"node-resolver-bkb8p\" (UID: \"e2ca5e21-6366-493a-b14b-b71baa6b54d7\") " pod="openshift-dns/node-resolver-bkb8p" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981672 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981718 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-hostroot\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981783 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-netns\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981755 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-netns\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981783 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-socket-dir-parent\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981928 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-openvswitch\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981974 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-ovn-kubernetes\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.981997 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-script-lib\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982020 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-etc-kubernetes\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982043 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-openvswitch\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982100 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-etc-kubernetes\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982049 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-os-release\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982055 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-ovn-kubernetes\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982165 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/354ebd50-7a67-481d-86db-4b3bf4753161-rootfs\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982214 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/354ebd50-7a67-481d-86db-4b3bf4753161-rootfs\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982272 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-systemd\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982219 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-systemd\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982331 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-log-socket\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982350 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982360 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-var-lib-cni-bin\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982399 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-log-socket\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982380 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-var-lib-cni-bin\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982416 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-daemon-config\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982436 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ed80eb00-6567-4878-91c2-979bb8e2f75c-os-release\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982450 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-system-cni-dir\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982483 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-run-netns\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982479 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-config\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982504 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-var-lib-kubelet\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982531 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-run-netns\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982508 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-system-cni-dir\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982530 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982568 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982580 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ed80eb00-6567-4878-91c2-979bb8e2f75c-cni-binary-copy\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982579 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-var-lib-kubelet\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982603 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/354ebd50-7a67-481d-86db-4b3bf4753161-proxy-tls\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982626 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-node-log\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982647 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-netd\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982666 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-cnibin\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982689 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-kubelet\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982707 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-ovn\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982728 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovn-node-metrics-cert\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982739 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-netd\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982751 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-run-k8s-cni-cncf-io\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982761 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-node-log\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982780 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-run-k8s-cni-cncf-io\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982803 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-ovn\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982823 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-kubelet\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982884 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq4cv\" (UniqueName: \"kubernetes.io/projected/e2ca5e21-6366-493a-b14b-b71baa6b54d7-kube-api-access-pq4cv\") pod \"node-resolver-bkb8p\" (UID: \"e2ca5e21-6366-493a-b14b-b71baa6b54d7\") " pod="openshift-dns/node-resolver-bkb8p" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982930 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8djp\" (UniqueName: \"kubernetes.io/projected/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-kube-api-access-r8djp\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982952 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-cnibin\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.982966 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5ttq\" (UniqueName: \"kubernetes.io/projected/ed80eb00-6567-4878-91c2-979bb8e2f75c-kube-api-access-s5ttq\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983010 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7980021c-dde3-4c14-a7b6-fbcc947a183d-cni-binary-copy\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983044 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-etc-openvswitch\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983076 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-bin\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983100 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-cni-dir\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983125 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ed80eb00-6567-4878-91c2-979bb8e2f75c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983149 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-slash\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983172 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-systemd-units\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983197 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxr2h\" (UniqueName: \"kubernetes.io/projected/354ebd50-7a67-481d-86db-4b3bf4753161-kube-api-access-gxr2h\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/354ebd50-7a67-481d-86db-4b3bf4753161-mcd-auth-proxy-config\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983231 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-systemd-units\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983247 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-var-lib-cni-multus\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983193 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-daemon-config\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983284 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-slash\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983291 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-run-multus-certs\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983332 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-run-multus-certs\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983237 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-bin\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983370 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz6b6\" (UniqueName: \"kubernetes.io/projected/7980021c-dde3-4c14-a7b6-fbcc947a183d-kube-api-access-jz6b6\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983379 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-host-var-lib-cni-multus\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983416 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ed80eb00-6567-4878-91c2-979bb8e2f75c-cni-binary-copy\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983421 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-os-release\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983442 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-cni-dir\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983463 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-conf-dir\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983475 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-os-release\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983485 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-env-overrides\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983504 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7980021c-dde3-4c14-a7b6-fbcc947a183d-multus-conf-dir\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.983925 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ed80eb00-6567-4878-91c2-979bb8e2f75c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.984189 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7980021c-dde3-4c14-a7b6-fbcc947a183d-cni-binary-copy\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.984216 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/354ebd50-7a67-481d-86db-4b3bf4753161-mcd-auth-proxy-config\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.984289 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-etc-openvswitch\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.984686 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-env-overrides\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.984811 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-script-lib\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.987460 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.989586 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovn-node-metrics-cert\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.989633 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/354ebd50-7a67-481d-86db-4b3bf4753161-proxy-tls\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:56 crc kubenswrapper[4799]: I1124 06:47:56.998927 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8djp\" (UniqueName: \"kubernetes.io/projected/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-kube-api-access-r8djp\") pod \"ovnkube-node-b7nd7\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.000249 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxr2h\" (UniqueName: \"kubernetes.io/projected/354ebd50-7a67-481d-86db-4b3bf4753161-kube-api-access-gxr2h\") pod \"machine-config-daemon-rzk7n\" (UID: \"354ebd50-7a67-481d-86db-4b3bf4753161\") " pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.001923 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.006766 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz6b6\" (UniqueName: \"kubernetes.io/projected/7980021c-dde3-4c14-a7b6-fbcc947a183d-kube-api-access-jz6b6\") pod \"multus-zvd7f\" (UID: \"7980021c-dde3-4c14-a7b6-fbcc947a183d\") " pod="openshift-multus/multus-zvd7f" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.008492 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq4cv\" (UniqueName: \"kubernetes.io/projected/e2ca5e21-6366-493a-b14b-b71baa6b54d7-kube-api-access-pq4cv\") pod \"node-resolver-bkb8p\" (UID: \"e2ca5e21-6366-493a-b14b-b71baa6b54d7\") " pod="openshift-dns/node-resolver-bkb8p" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.008822 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5ttq\" (UniqueName: \"kubernetes.io/projected/ed80eb00-6567-4878-91c2-979bb8e2f75c-kube-api-access-s5ttq\") pod \"multus-additional-cni-plugins-zzvdv\" (UID: \"ed80eb00-6567-4878-91c2-979bb8e2f75c\") " pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.018666 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.031324 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.043346 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.054464 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.066435 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.162168 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bkb8p" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.179524 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.192055 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zvd7f" Nov 24 06:47:57 crc kubenswrapper[4799]: W1124 06:47:57.193065 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded80eb00_6567_4878_91c2_979bb8e2f75c.slice/crio-195e9057195a17e352fb16d62cff3baafee847f7e8c8fa2655fe9cda8e12b825 WatchSource:0}: Error finding container 195e9057195a17e352fb16d62cff3baafee847f7e8c8fa2655fe9cda8e12b825: Status 404 returned error can't find the container with id 195e9057195a17e352fb16d62cff3baafee847f7e8c8fa2655fe9cda8e12b825 Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.213594 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.215378 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.286390 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.286620 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:47:59.286599969 +0000 UTC m=+24.942582443 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.387792 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.387868 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.387930 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.387960 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388042 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388074 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388087 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388100 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388144 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:59.38812394 +0000 UTC m=+25.044106414 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388139 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388182 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388197 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388257 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:59.388232553 +0000 UTC m=+25.044215027 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388063 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388384 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:59.388372477 +0000 UTC m=+25.044354951 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.388981 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:47:59.388968454 +0000 UTC m=+25.044950928 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.628126 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.628161 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.628199 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.628296 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.628708 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:47:57 crc kubenswrapper[4799]: E1124 06:47:57.628831 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.632887 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.633741 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.634627 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.635460 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.637320 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.638064 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.639359 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.640107 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.641510 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.642240 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.643589 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.644590 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.645833 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.646483 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.647026 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.649463 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.650249 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.651291 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.651883 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.652534 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.653436 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.654004 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.654528 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.655734 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.656312 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.657339 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.657987 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.658901 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.659973 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.660610 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.661192 4799 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.661317 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.663842 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.665109 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.665557 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.667569 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.669165 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.670358 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.672120 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.672900 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.673950 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.674620 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.675660 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.676312 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.677200 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.677791 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.678792 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.679537 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.680583 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.681118 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.681961 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.682520 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.683213 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.684213 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.787941 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488"} Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.788036 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9"} Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.788050 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"08a0827e0950b94175ee67c68d7981dcfd35d5cddd12b331d5e15bcc7c04abe0"} Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.789398 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" event={"ID":"ed80eb00-6567-4878-91c2-979bb8e2f75c","Type":"ContainerStarted","Data":"496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e"} Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.789425 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" event={"ID":"ed80eb00-6567-4878-91c2-979bb8e2f75c","Type":"ContainerStarted","Data":"195e9057195a17e352fb16d62cff3baafee847f7e8c8fa2655fe9cda8e12b825"} Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.790678 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bkb8p" event={"ID":"e2ca5e21-6366-493a-b14b-b71baa6b54d7","Type":"ContainerStarted","Data":"c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95"} Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.790713 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bkb8p" event={"ID":"e2ca5e21-6366-493a-b14b-b71baa6b54d7","Type":"ContainerStarted","Data":"69c61fefc0d3c1136f675f2b2ca73ef7f04d1504fad2e2fa0e5315a5274fdfbc"} Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.792200 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c" exitCode=0 Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.792287 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c"} Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.792356 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"ed853cc5cdad54443adb0e29e0a8965d4896c34402e739666d0b1d156d04aa85"} Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.794069 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvd7f" event={"ID":"7980021c-dde3-4c14-a7b6-fbcc947a183d","Type":"ContainerStarted","Data":"ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d"} Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.794097 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvd7f" event={"ID":"7980021c-dde3-4c14-a7b6-fbcc947a183d","Type":"ContainerStarted","Data":"de213ecbd831749f748da47d0b573076988661af827b48dbd5e8cea92a5703e3"} Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.813057 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.829809 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.847155 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.891043 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.920613 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.932362 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.956828 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.972447 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:57 crc kubenswrapper[4799]: I1124 06:47:57.990103 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.003091 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.018898 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.036122 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.041282 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.055713 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.060397 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.061498 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.071928 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.088112 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.105168 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.124927 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.139578 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.156912 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.170918 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.192136 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.205088 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.223033 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.246095 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.273034 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.296016 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.317287 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.337420 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.354740 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.371094 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.388004 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.403388 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.419102 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.445405 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.488294 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.533095 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.558334 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.575282 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.590767 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.608351 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.806846 4799 generic.go:334] "Generic (PLEG): container finished" podID="ed80eb00-6567-4878-91c2-979bb8e2f75c" containerID="496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e" exitCode=0 Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.806967 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" event={"ID":"ed80eb00-6567-4878-91c2-979bb8e2f75c","Type":"ContainerDied","Data":"496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e"} Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.810221 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938"} Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.817122 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95"} Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.817196 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1"} Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.817213 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a"} Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.817228 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74"} Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.817241 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a"} Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.817255 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a"} Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.841562 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.857011 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.875338 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.893835 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.913790 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.928608 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.946651 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.963675 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:58 crc kubenswrapper[4799]: I1124 06:47:58.993909 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:58Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.011569 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.024324 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.036221 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.047159 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.060961 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.073410 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.097611 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.137617 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.177926 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.223399 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.256898 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.295814 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.311318 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.311562 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:48:03.311527156 +0000 UTC m=+28.967509630 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.338534 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.378813 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.412348 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.412414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.412450 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.412482 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.412585 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.412670 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:03.412635485 +0000 UTC m=+29.068617979 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.413033 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.413137 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:03.413113009 +0000 UTC m=+29.069095483 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.413033 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.413348 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.413445 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.413596 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:03.413569972 +0000 UTC m=+29.069552456 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.413742 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.413830 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.413934 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.414055 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:03.414042035 +0000 UTC m=+29.070024519 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.420558 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.459544 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.501393 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.538383 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.586968 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.627894 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.628090 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.627929 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.628806 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.628936 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:47:59 crc kubenswrapper[4799]: E1124 06:47:59.629198 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.822718 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" event={"ID":"ed80eb00-6567-4878-91c2-979bb8e2f75c","Type":"ContainerStarted","Data":"50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e"} Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.847936 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.869940 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.883838 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.900219 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.911598 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.926493 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.939839 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.961021 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.975321 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:47:59 crc kubenswrapper[4799]: I1124 06:47:59.990568 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:47:59Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.022743 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.057514 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.096695 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.138382 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.827681 4799 generic.go:334] "Generic (PLEG): container finished" podID="ed80eb00-6567-4878-91c2-979bb8e2f75c" containerID="50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e" exitCode=0 Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.828012 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" event={"ID":"ed80eb00-6567-4878-91c2-979bb8e2f75c","Type":"ContainerDied","Data":"50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e"} Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.843939 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54"} Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.846260 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.864123 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.883272 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.901788 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.919443 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.939064 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.953523 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.968833 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:00 crc kubenswrapper[4799]: I1124 06:48:00.997534 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:00Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.014577 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.027814 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.040731 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.055462 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.084931 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.627614 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.627751 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:01 crc kubenswrapper[4799]: E1124 06:48:01.627777 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.627829 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:01 crc kubenswrapper[4799]: E1124 06:48:01.627947 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:01 crc kubenswrapper[4799]: E1124 06:48:01.628068 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.852924 4799 generic.go:334] "Generic (PLEG): container finished" podID="ed80eb00-6567-4878-91c2-979bb8e2f75c" containerID="6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe" exitCode=0 Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.852989 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" event={"ID":"ed80eb00-6567-4878-91c2-979bb8e2f75c","Type":"ContainerDied","Data":"6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe"} Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.874398 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.911432 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.930947 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.947822 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.965382 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.983731 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:01 crc kubenswrapper[4799]: I1124 06:48:01.999536 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:01Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.013463 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.024417 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.028694 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.028744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.028780 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.028897 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.029604 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.037739 4799 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.038123 4799 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.039429 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.039499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.039517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.039542 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.039559 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.041501 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: E1124 06:48:02.054567 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.059685 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.060076 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.060191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.060329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.060422 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.062511 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.078999 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: E1124 06:48:02.079540 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.086928 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.086980 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.086996 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.087019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.087035 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.095890 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: E1124 06:48:02.105473 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.110180 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.111457 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.111477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.111485 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.111500 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.111510 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: E1124 06:48:02.131062 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.136255 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.136302 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.136312 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.136329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.136340 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: E1124 06:48:02.150055 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: E1124 06:48:02.150177 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.153329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.153381 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.153393 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.153410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.153425 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.256553 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.256589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.256597 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.256613 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.256621 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.359640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.359678 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.359687 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.359702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.359711 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.462915 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.462985 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.463005 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.463041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.463064 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.566339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.566416 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.566431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.566450 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.566462 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.668506 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.668577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.668594 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.668613 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.668625 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.772032 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.772097 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.772110 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.772133 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.772146 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.859595 4799 generic.go:334] "Generic (PLEG): container finished" podID="ed80eb00-6567-4878-91c2-979bb8e2f75c" containerID="e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44" exitCode=0 Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.859673 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" event={"ID":"ed80eb00-6567-4878-91c2-979bb8e2f75c","Type":"ContainerDied","Data":"e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.877753 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.878070 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.878099 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.878108 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.878124 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.878285 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.895306 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-fzb2l"] Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.896037 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fzb2l" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.901261 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.902113 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.902588 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.902659 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.905746 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.932693 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.948570 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.967444 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmxgw\" (UniqueName: \"kubernetes.io/projected/75f2af19-8c5a-454d-a9b1-7f9ae60a9e48-kube-api-access-mmxgw\") pod \"node-ca-fzb2l\" (UID: \"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\") " pod="openshift-image-registry/node-ca-fzb2l" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.967494 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/75f2af19-8c5a-454d-a9b1-7f9ae60a9e48-serviceca\") pod \"node-ca-fzb2l\" (UID: \"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\") " pod="openshift-image-registry/node-ca-fzb2l" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.967557 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/75f2af19-8c5a-454d-a9b1-7f9ae60a9e48-host\") pod \"node-ca-fzb2l\" (UID: \"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\") " pod="openshift-image-registry/node-ca-fzb2l" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.969011 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.981963 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.982009 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.982026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.982052 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.982072 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:02Z","lastTransitionTime":"2025-11-24T06:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:02 crc kubenswrapper[4799]: I1124 06:48:02.986788 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.002963 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.019750 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.040972 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.059651 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.068190 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/75f2af19-8c5a-454d-a9b1-7f9ae60a9e48-host\") pod \"node-ca-fzb2l\" (UID: \"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\") " pod="openshift-image-registry/node-ca-fzb2l" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.068267 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxgw\" (UniqueName: \"kubernetes.io/projected/75f2af19-8c5a-454d-a9b1-7f9ae60a9e48-kube-api-access-mmxgw\") pod \"node-ca-fzb2l\" (UID: \"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\") " pod="openshift-image-registry/node-ca-fzb2l" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.068296 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/75f2af19-8c5a-454d-a9b1-7f9ae60a9e48-serviceca\") pod \"node-ca-fzb2l\" (UID: \"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\") " pod="openshift-image-registry/node-ca-fzb2l" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.068412 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/75f2af19-8c5a-454d-a9b1-7f9ae60a9e48-host\") pod \"node-ca-fzb2l\" (UID: \"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\") " pod="openshift-image-registry/node-ca-fzb2l" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.073488 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/75f2af19-8c5a-454d-a9b1-7f9ae60a9e48-serviceca\") pod \"node-ca-fzb2l\" (UID: \"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\") " pod="openshift-image-registry/node-ca-fzb2l" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.084525 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.085195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.085229 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.085239 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.085254 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.085264 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:03Z","lastTransitionTime":"2025-11-24T06:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.089489 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmxgw\" (UniqueName: \"kubernetes.io/projected/75f2af19-8c5a-454d-a9b1-7f9ae60a9e48-kube-api-access-mmxgw\") pod \"node-ca-fzb2l\" (UID: \"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\") " pod="openshift-image-registry/node-ca-fzb2l" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.099981 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.115668 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.131318 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.153202 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.167172 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.188137 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.188421 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.188453 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.188472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.188497 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.188514 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:03Z","lastTransitionTime":"2025-11-24T06:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.207170 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.228148 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fzb2l" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.228884 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.248127 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.274342 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.296297 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.298912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.299007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.299031 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.299061 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.299081 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:03Z","lastTransitionTime":"2025-11-24T06:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.328232 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.362281 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.375450 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.375688 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:48:11.375653763 +0000 UTC m=+37.031636307 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.377979 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.392198 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.406306 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.407246 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.407307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.407322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.407350 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.407367 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:03Z","lastTransitionTime":"2025-11-24T06:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.419943 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.436169 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.476630 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.476703 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.476741 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.476769 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.476930 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.477024 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:11.476990256 +0000 UTC m=+37.132972740 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.477385 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.477551 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.477578 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.477595 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.477443 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.477627 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:11.477617374 +0000 UTC m=+37.133599858 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.477629 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.477674 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:11.477666465 +0000 UTC m=+37.133648949 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.477703 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.477862 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:11.477795479 +0000 UTC m=+37.133778133 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.511525 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.511574 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.511588 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.511607 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.511619 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:03Z","lastTransitionTime":"2025-11-24T06:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.614345 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.614407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.614420 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.614442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.614458 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:03Z","lastTransitionTime":"2025-11-24T06:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.627660 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.627747 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.627816 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.627862 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.627905 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:03 crc kubenswrapper[4799]: E1124 06:48:03.628405 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.717450 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.717521 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.717546 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.717585 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.717687 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:03Z","lastTransitionTime":"2025-11-24T06:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.820729 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.820785 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.820799 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.820819 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.820835 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:03Z","lastTransitionTime":"2025-11-24T06:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.877138 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" event={"ID":"ed80eb00-6567-4878-91c2-979bb8e2f75c","Type":"ContainerStarted","Data":"3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.881165 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fzb2l" event={"ID":"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48","Type":"ContainerStarted","Data":"292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.881275 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fzb2l" event={"ID":"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48","Type":"ContainerStarted","Data":"b27fc8effb98757617db04093871df31445c1f178cd8d80c4bfff0506646fd6f"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.890069 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.890683 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.890885 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.903685 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.923441 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.923867 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.923895 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.923924 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.923942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.923955 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:03Z","lastTransitionTime":"2025-11-24T06:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.962308 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.969555 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.971749 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.985690 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:03 crc kubenswrapper[4799]: I1124 06:48:03.999192 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.015618 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.027280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.027328 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.027342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.027376 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.027391 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:04Z","lastTransitionTime":"2025-11-24T06:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.030446 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.050328 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.068534 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.083539 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.098733 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.111946 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.127466 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.130200 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.130239 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.130250 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.130265 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.130277 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:04Z","lastTransitionTime":"2025-11-24T06:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.144374 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.170395 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.185275 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.198993 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.211021 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.229052 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.233898 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.233938 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.233954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.233970 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.233984 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:04Z","lastTransitionTime":"2025-11-24T06:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.246047 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.268611 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.287730 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.306426 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.335141 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.338041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.338086 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.338100 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.338124 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.338140 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:04Z","lastTransitionTime":"2025-11-24T06:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.353434 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.371126 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.390086 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.412170 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.428785 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.440552 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.440598 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.440609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.440626 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.440637 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:04Z","lastTransitionTime":"2025-11-24T06:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.453829 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.543887 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.543959 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.543974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.543997 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.544011 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:04Z","lastTransitionTime":"2025-11-24T06:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.647715 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.647757 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.647767 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.647783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.647808 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:04Z","lastTransitionTime":"2025-11-24T06:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.751328 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.751373 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.751385 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.751404 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.751418 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:04Z","lastTransitionTime":"2025-11-24T06:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.854993 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.855041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.855053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.855074 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.855089 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:04Z","lastTransitionTime":"2025-11-24T06:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.899129 4799 generic.go:334] "Generic (PLEG): container finished" podID="ed80eb00-6567-4878-91c2-979bb8e2f75c" containerID="3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c" exitCode=0 Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.899204 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" event={"ID":"ed80eb00-6567-4878-91c2-979bb8e2f75c","Type":"ContainerDied","Data":"3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c"} Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.899385 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.920060 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.934947 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.951013 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.958455 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.958494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.958505 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.958520 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.958531 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:04Z","lastTransitionTime":"2025-11-24T06:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.970928 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:04 crc kubenswrapper[4799]: I1124 06:48:04.985223 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:04Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.004427 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.017904 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.042026 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.061464 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.061623 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.061701 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.061798 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.061827 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:05Z","lastTransitionTime":"2025-11-24T06:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.071117 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.097996 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.114970 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.128261 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.141615 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.156693 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.167338 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.167378 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.167392 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.167410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.167419 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:05Z","lastTransitionTime":"2025-11-24T06:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.174035 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.270527 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.270582 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.270594 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.270614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.270632 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:05Z","lastTransitionTime":"2025-11-24T06:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.377143 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.377204 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.377217 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.377234 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.377243 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:05Z","lastTransitionTime":"2025-11-24T06:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.480420 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.480450 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.480460 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.480473 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.480482 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:05Z","lastTransitionTime":"2025-11-24T06:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.583776 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.583824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.583835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.583870 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.583882 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:05Z","lastTransitionTime":"2025-11-24T06:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.627685 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:05 crc kubenswrapper[4799]: E1124 06:48:05.627811 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.627978 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:05 crc kubenswrapper[4799]: E1124 06:48:05.628236 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.628409 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:05 crc kubenswrapper[4799]: E1124 06:48:05.628593 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.651509 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.667225 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.680430 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.686274 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.686309 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.686318 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.686333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.686342 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:05Z","lastTransitionTime":"2025-11-24T06:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.698707 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.718974 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.736253 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.750211 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.771336 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.784448 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.790920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.790957 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.790967 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.790985 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.790996 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:05Z","lastTransitionTime":"2025-11-24T06:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.810083 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.827783 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.844176 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.857462 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.868075 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.880220 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.893358 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.893388 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.893399 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.893414 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.893425 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:05Z","lastTransitionTime":"2025-11-24T06:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.905365 4799 generic.go:334] "Generic (PLEG): container finished" podID="ed80eb00-6567-4878-91c2-979bb8e2f75c" containerID="4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8" exitCode=0 Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.905461 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" event={"ID":"ed80eb00-6567-4878-91c2-979bb8e2f75c","Type":"ContainerDied","Data":"4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8"} Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.905558 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.924798 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.936350 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.960766 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.977742 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.993273 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.996431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.996475 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.996493 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.996516 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:05 crc kubenswrapper[4799]: I1124 06:48:05.996529 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:05Z","lastTransitionTime":"2025-11-24T06:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.009163 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.071591 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.088026 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.099157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.099198 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.099210 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.099228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.099241 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:06Z","lastTransitionTime":"2025-11-24T06:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.102623 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.121917 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.139985 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.155036 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.166946 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.178609 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.191734 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.201643 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.201692 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.201706 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.201722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.201734 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:06Z","lastTransitionTime":"2025-11-24T06:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.304798 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.305173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.305258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.305329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.305389 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:06Z","lastTransitionTime":"2025-11-24T06:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.408342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.408723 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.408834 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.408991 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.409100 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:06Z","lastTransitionTime":"2025-11-24T06:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.512143 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.512195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.512210 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.512226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.512239 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:06Z","lastTransitionTime":"2025-11-24T06:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.614553 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.614607 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.614624 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.614644 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.614658 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:06Z","lastTransitionTime":"2025-11-24T06:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.717100 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.717136 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.717147 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.717160 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.717169 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:06Z","lastTransitionTime":"2025-11-24T06:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.819662 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.819704 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.819713 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.819752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.819765 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:06Z","lastTransitionTime":"2025-11-24T06:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.912933 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" event={"ID":"ed80eb00-6567-4878-91c2-979bb8e2f75c","Type":"ContainerStarted","Data":"01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111"} Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.921791 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.921875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.921888 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.921902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.921912 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:06Z","lastTransitionTime":"2025-11-24T06:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.928841 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.950060 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.965957 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.980464 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:06 crc kubenswrapper[4799]: I1124 06:48:06.995000 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.007821 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.024592 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.024642 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.024668 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.024685 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.024699 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:07Z","lastTransitionTime":"2025-11-24T06:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.028240 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.048213 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.076458 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.096791 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.125209 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.128783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.128826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.128838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.128874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.128886 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:07Z","lastTransitionTime":"2025-11-24T06:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.139159 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.149253 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.163947 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.178538 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.231333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.231357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.231366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.231379 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.231390 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:07Z","lastTransitionTime":"2025-11-24T06:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.334227 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.334280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.334294 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.334316 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.334332 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:07Z","lastTransitionTime":"2025-11-24T06:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.437896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.437954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.437965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.437983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.437993 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:07Z","lastTransitionTime":"2025-11-24T06:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.541931 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.541982 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.541996 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.542013 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.542023 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:07Z","lastTransitionTime":"2025-11-24T06:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.627440 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.627521 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.627557 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:07 crc kubenswrapper[4799]: E1124 06:48:07.627629 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:07 crc kubenswrapper[4799]: E1124 06:48:07.627754 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:07 crc kubenswrapper[4799]: E1124 06:48:07.627924 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.646193 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.646260 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.646281 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.646308 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.646331 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:07Z","lastTransitionTime":"2025-11-24T06:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.750631 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.750689 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.750700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.750721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.750732 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:07Z","lastTransitionTime":"2025-11-24T06:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.854168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.854309 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.854340 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.854387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.854419 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:07Z","lastTransitionTime":"2025-11-24T06:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.920640 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/0.log" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.927567 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66" exitCode=1 Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.927628 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66"} Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.929158 4799 scope.go:117] "RemoveContainer" containerID="5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.943794 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.961164 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.961213 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.961225 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.961243 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.961260 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:07Z","lastTransitionTime":"2025-11-24T06:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.963837 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:07 crc kubenswrapper[4799]: I1124 06:48:07.983227 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.000445 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:07Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.017634 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.041574 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.056753 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.064362 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.064411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.064426 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.064444 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.064456 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:08Z","lastTransitionTime":"2025-11-24T06:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.072531 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.085815 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.105622 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.124382 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.143429 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.168415 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.200291 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.200396 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.200411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.200437 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.200458 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:08Z","lastTransitionTime":"2025-11-24T06:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.247265 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.264648 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:07Z\\\",\\\"message\\\":\\\"/informers/externalversions/factory.go:141\\\\nI1124 06:48:07.010258 6013 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:07.010551 6013 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:07.010744 6013 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:07.011328 6013 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 06:48:07.011385 6013 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 06:48:07.011409 6013 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 06:48:07.011415 6013 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:07.011430 6013 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 06:48:07.011453 6013 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:07.011454 6013 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 06:48:07.011484 6013 factory.go:656] Stopping watch factory\\\\nI1124 06:48:07.011509 6013 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.304020 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.304068 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.304082 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.304101 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.304114 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:08Z","lastTransitionTime":"2025-11-24T06:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.408729 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.408794 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.408812 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.408838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.408886 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:08Z","lastTransitionTime":"2025-11-24T06:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.532767 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.533108 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.533211 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.533335 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.533449 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:08Z","lastTransitionTime":"2025-11-24T06:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.636580 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.636634 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.636649 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.636671 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.636686 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:08Z","lastTransitionTime":"2025-11-24T06:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.739279 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.739333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.739343 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.739364 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.739377 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:08Z","lastTransitionTime":"2025-11-24T06:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.842327 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.842384 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.842402 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.842434 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.842446 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:08Z","lastTransitionTime":"2025-11-24T06:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.933307 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/0.log" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.938328 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe"} Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.938510 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.944361 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.944388 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.944397 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.944412 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.944423 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:08Z","lastTransitionTime":"2025-11-24T06:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.961187 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.972183 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:08 crc kubenswrapper[4799]: I1124 06:48:08.987823 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.001751 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:08Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.016562 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.031781 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.046263 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.047440 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.047507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.047538 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.047564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.047581 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:09Z","lastTransitionTime":"2025-11-24T06:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.061051 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.079007 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:07Z\\\",\\\"message\\\":\\\"/informers/externalversions/factory.go:141\\\\nI1124 06:48:07.010258 6013 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:07.010551 6013 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:07.010744 6013 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:07.011328 6013 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 06:48:07.011385 6013 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 06:48:07.011409 6013 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 06:48:07.011415 6013 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:07.011430 6013 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 06:48:07.011453 6013 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:07.011454 6013 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 06:48:07.011484 6013 factory.go:656] Stopping watch factory\\\\nI1124 06:48:07.011509 6013 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.093357 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.105977 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.120885 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.132664 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.146298 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.150950 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.151012 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.151025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.151045 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.151060 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:09Z","lastTransitionTime":"2025-11-24T06:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.189367 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.251064 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.255006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.255084 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.255113 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.255163 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.255190 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:09Z","lastTransitionTime":"2025-11-24T06:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.357978 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.358022 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.358036 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.358059 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.358073 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:09Z","lastTransitionTime":"2025-11-24T06:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.460782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.460875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.460896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.460923 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.460940 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:09Z","lastTransitionTime":"2025-11-24T06:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.565571 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.565671 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.565696 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.565743 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.565776 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:09Z","lastTransitionTime":"2025-11-24T06:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.628342 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.628407 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.628378 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:09 crc kubenswrapper[4799]: E1124 06:48:09.628589 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:09 crc kubenswrapper[4799]: E1124 06:48:09.628818 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:09 crc kubenswrapper[4799]: E1124 06:48:09.629005 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.669255 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.669313 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.669325 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.669342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.669353 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:09Z","lastTransitionTime":"2025-11-24T06:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.773186 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.773236 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.773487 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.773520 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.773531 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:09Z","lastTransitionTime":"2025-11-24T06:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.781835 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k"] Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.782827 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.787394 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.787666 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.798290 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.814651 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.830591 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.849307 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.862656 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.863871 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xktbj\" (UniqueName: \"kubernetes.io/projected/c56a6245-1460-4588-b33c-188991e12266-kube-api-access-xktbj\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.863946 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c56a6245-1460-4588-b33c-188991e12266-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.864047 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c56a6245-1460-4588-b33c-188991e12266-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.864099 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c56a6245-1460-4588-b33c-188991e12266-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.876814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.876896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.876912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.876934 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.876951 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:09Z","lastTransitionTime":"2025-11-24T06:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.877464 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.910592 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.928354 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.945455 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/1.log" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.946451 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/0.log" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.950080 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe" exitCode=1 Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.950156 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe"} Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.950219 4799 scope.go:117] "RemoveContainer" containerID="5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.952041 4799 scope.go:117] "RemoveContainer" containerID="eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe" Nov 24 06:48:09 crc kubenswrapper[4799]: E1124 06:48:09.952755 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.953789 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.966499 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c56a6245-1460-4588-b33c-188991e12266-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.966768 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c56a6245-1460-4588-b33c-188991e12266-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.967163 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c56a6245-1460-4588-b33c-188991e12266-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.967297 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.967618 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c56a6245-1460-4588-b33c-188991e12266-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.967837 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xktbj\" (UniqueName: \"kubernetes.io/projected/c56a6245-1460-4588-b33c-188991e12266-kube-api-access-xktbj\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.967951 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c56a6245-1460-4588-b33c-188991e12266-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.978641 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c56a6245-1460-4588-b33c-188991e12266-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.979564 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.980152 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.980182 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.980194 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.980213 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.980229 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:09Z","lastTransitionTime":"2025-11-24T06:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.988623 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.991489 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xktbj\" (UniqueName: \"kubernetes.io/projected/c56a6245-1460-4588-b33c-188991e12266-kube-api-access-xktbj\") pod \"ovnkube-control-plane-749d76644c-2vq5k\" (UID: \"c56a6245-1460-4588-b33c-188991e12266\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:09 crc kubenswrapper[4799]: I1124 06:48:09.996618 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.009917 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.023490 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.036564 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.055415 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:07Z\\\",\\\"message\\\":\\\"/informers/externalversions/factory.go:141\\\\nI1124 06:48:07.010258 6013 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:07.010551 6013 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:07.010744 6013 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:07.011328 6013 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 06:48:07.011385 6013 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 06:48:07.011409 6013 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 06:48:07.011415 6013 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:07.011430 6013 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 06:48:07.011453 6013 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:07.011454 6013 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 06:48:07.011484 6013 factory.go:656] Stopping watch factory\\\\nI1124 06:48:07.011509 6013 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.066345 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.083187 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.083670 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.083732 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.083759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.083775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.083784 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:10Z","lastTransitionTime":"2025-11-24T06:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.101090 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.106770 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.113690 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: W1124 06:48:10.121615 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc56a6245_1460_4588_b33c_188991e12266.slice/crio-a277f20ed78458100e00d4d6343cac8349732ed91988875c3acacfa8f7460fba WatchSource:0}: Error finding container a277f20ed78458100e00d4d6343cac8349732ed91988875c3acacfa8f7460fba: Status 404 returned error can't find the container with id a277f20ed78458100e00d4d6343cac8349732ed91988875c3acacfa8f7460fba Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.131198 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.144655 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.173377 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.186671 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.186703 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.186714 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.186729 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.186741 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:10Z","lastTransitionTime":"2025-11-24T06:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.189010 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.203303 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.214584 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.232728 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.250474 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.266433 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.282666 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.288811 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.288839 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.288862 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.288875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.288884 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:10Z","lastTransitionTime":"2025-11-24T06:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.304350 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.325762 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5509d14f119f7909712d22cb6a4225117694bc244e71eae9802160a953e86a66\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:07Z\\\",\\\"message\\\":\\\"/informers/externalversions/factory.go:141\\\\nI1124 06:48:07.010258 6013 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:07.010551 6013 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:07.010744 6013 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:07.011328 6013 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 06:48:07.011385 6013 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 06:48:07.011409 6013 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 06:48:07.011415 6013 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:07.011430 6013 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 06:48:07.011453 6013 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:07.011454 6013 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 06:48:07.011484 6013 factory.go:656] Stopping watch factory\\\\nI1124 06:48:07.011509 6013 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"173564 6243 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1124 06:48:09.173566 6243 lb_config.go:1031] Cluster endpoints for openshift-dns/dns-default for network=default are: map[]\\\\nI1124 06:48:09.173571 6243 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1124 06:48:09.173202 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:48:09.173585 6243 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-ch\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.391809 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.391890 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.391909 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.391929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.391945 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:10Z","lastTransitionTime":"2025-11-24T06:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.494691 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.494982 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.495058 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.495203 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.495281 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:10Z","lastTransitionTime":"2025-11-24T06:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.598657 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.598737 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.598757 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.598785 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.598805 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:10Z","lastTransitionTime":"2025-11-24T06:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.701647 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.701681 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.701688 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.701702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.701711 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:10Z","lastTransitionTime":"2025-11-24T06:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.803830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.803929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.803948 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.803970 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.803986 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:10Z","lastTransitionTime":"2025-11-24T06:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.906992 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.907037 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.907047 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.907064 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.907077 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:10Z","lastTransitionTime":"2025-11-24T06:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.955247 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/1.log" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.958926 4799 scope.go:117] "RemoveContainer" containerID="eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe" Nov 24 06:48:10 crc kubenswrapper[4799]: E1124 06:48:10.959097 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.960724 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" event={"ID":"c56a6245-1460-4588-b33c-188991e12266","Type":"ContainerStarted","Data":"c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.960793 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" event={"ID":"c56a6245-1460-4588-b33c-188991e12266","Type":"ContainerStarted","Data":"711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.960819 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" event={"ID":"c56a6245-1460-4588-b33c-188991e12266","Type":"ContainerStarted","Data":"a277f20ed78458100e00d4d6343cac8349732ed91988875c3acacfa8f7460fba"} Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.973580 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:10 crc kubenswrapper[4799]: I1124 06:48:10.987590 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:10Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.007398 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.009290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.009322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.009334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.009352 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.009367 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:11Z","lastTransitionTime":"2025-11-24T06:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.023600 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.041137 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.054988 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.075719 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.093380 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.107019 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.116892 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.116951 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.116965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.116983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.117012 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:11Z","lastTransitionTime":"2025-11-24T06:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.123041 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.138857 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.152878 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.167188 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.187072 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.198754 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.220263 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.220550 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.220630 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.220759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.220889 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:11Z","lastTransitionTime":"2025-11-24T06:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.227590 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"173564 6243 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1124 06:48:09.173566 6243 lb_config.go:1031] Cluster endpoints for openshift-dns/dns-default for network=default are: map[]\\\\nI1124 06:48:09.173571 6243 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1124 06:48:09.173202 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:48:09.173585 6243 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-ch\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.241779 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.261775 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.281502 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.286512 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-ldbr4"] Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.287131 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.287209 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.298530 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.315422 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.323464 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.323510 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.323521 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.323536 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.323547 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:11Z","lastTransitionTime":"2025-11-24T06:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.331801 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.346264 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.363075 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.382364 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.382498 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.382561 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:48:27.382526043 +0000 UTC m=+53.038508537 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.382721 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c87kr\" (UniqueName: \"kubernetes.io/projected/f587ff07-479b-4c95-95cb-406faffcfb68-kube-api-access-c87kr\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.387316 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"173564 6243 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1124 06:48:09.173566 6243 lb_config.go:1031] Cluster endpoints for openshift-dns/dns-default for network=default are: map[]\\\\nI1124 06:48:09.173571 6243 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1124 06:48:09.173202 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:48:09.173585 6243 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-ch\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.406569 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.420146 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.425931 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.426337 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.426534 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.426634 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.426714 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:11Z","lastTransitionTime":"2025-11-24T06:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.432216 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.443198 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.461641 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.480753 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.484032 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.484089 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c87kr\" (UniqueName: \"kubernetes.io/projected/f587ff07-479b-4c95-95cb-406faffcfb68-kube-api-access-c87kr\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.484141 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.484204 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.484250 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.484287 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.484398 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.484476 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.484516 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.484541 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.484547 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.484489 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:27.484463583 +0000 UTC m=+53.140446077 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.484760 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:27.48471825 +0000 UTC m=+53.140700914 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.484792 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs podName:f587ff07-479b-4c95-95cb-406faffcfb68 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:11.984775842 +0000 UTC m=+37.640758546 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs") pod "network-metrics-daemon-ldbr4" (UID: "f587ff07-479b-4c95-95cb-406faffcfb68") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.484807 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.485025 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:27.484975328 +0000 UTC m=+53.140957842 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.485330 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.485385 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.485400 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.485473 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:27.485452581 +0000 UTC m=+53.141435055 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.505873 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c87kr\" (UniqueName: \"kubernetes.io/projected/f587ff07-479b-4c95-95cb-406faffcfb68-kube-api-access-c87kr\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.517978 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.530346 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.530431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.530456 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.530487 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.530510 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:11Z","lastTransitionTime":"2025-11-24T06:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.533965 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.551297 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.570488 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.590609 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.606338 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.622868 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.627418 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.627501 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.627444 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.627584 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.627724 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.627814 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.632729 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.632784 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.632800 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.632823 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.632836 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:11Z","lastTransitionTime":"2025-11-24T06:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.639933 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.665457 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.681120 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.702778 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.713225 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.726365 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.735592 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.735627 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.735640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.735659 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.735672 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:11Z","lastTransitionTime":"2025-11-24T06:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.738013 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.751622 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.764248 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.777461 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.795441 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"173564 6243 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1124 06:48:09.173566 6243 lb_config.go:1031] Cluster endpoints for openshift-dns/dns-default for network=default are: map[]\\\\nI1124 06:48:09.173571 6243 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1124 06:48:09.173202 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:48:09.173585 6243 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-ch\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:11Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.839186 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.839231 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.839245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.839260 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.839273 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:11Z","lastTransitionTime":"2025-11-24T06:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.942961 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.943021 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.943031 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.943049 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.943060 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:11Z","lastTransitionTime":"2025-11-24T06:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:11 crc kubenswrapper[4799]: I1124 06:48:11.996766 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.996958 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:11 crc kubenswrapper[4799]: E1124 06:48:11.997052 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs podName:f587ff07-479b-4c95-95cb-406faffcfb68 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:12.997032835 +0000 UTC m=+38.653015309 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs") pod "network-metrics-daemon-ldbr4" (UID: "f587ff07-479b-4c95-95cb-406faffcfb68") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.046810 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.046888 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.046899 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.046912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.046921 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.150622 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.150675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.150687 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.150704 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.150717 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.251973 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.252058 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.252079 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.252113 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.252133 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: E1124 06:48:12.272051 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:12Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.276881 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.276939 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.276954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.276975 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.276988 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: E1124 06:48:12.297600 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:12Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.302291 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.302326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.302337 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.302353 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.302364 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: E1124 06:48:12.317750 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:12Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.322198 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.322236 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.322245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.322263 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.322274 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: E1124 06:48:12.336369 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:12Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.341702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.341760 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.341771 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.341790 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.341805 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: E1124 06:48:12.358319 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:12Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:12 crc kubenswrapper[4799]: E1124 06:48:12.358447 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.360279 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.360311 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.360322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.360343 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.360355 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.464070 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.464139 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.464160 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.464191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.464212 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.568277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.568373 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.568393 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.568424 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.568446 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.628319 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:12 crc kubenswrapper[4799]: E1124 06:48:12.628601 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.671904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.671974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.671992 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.672019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.672042 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.780971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.781039 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.781054 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.781075 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.781087 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.883972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.884038 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.884053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.884081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.884097 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.986553 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.986600 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.986610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.986671 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:12 crc kubenswrapper[4799]: I1124 06:48:12.986683 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:12Z","lastTransitionTime":"2025-11-24T06:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.008506 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:13 crc kubenswrapper[4799]: E1124 06:48:13.008783 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:13 crc kubenswrapper[4799]: E1124 06:48:13.008928 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs podName:f587ff07-479b-4c95-95cb-406faffcfb68 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:15.008898132 +0000 UTC m=+40.664880836 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs") pod "network-metrics-daemon-ldbr4" (UID: "f587ff07-479b-4c95-95cb-406faffcfb68") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.089653 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.089747 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.089774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.089808 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.089835 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:13Z","lastTransitionTime":"2025-11-24T06:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.193583 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.193670 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.193707 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.193731 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.193743 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:13Z","lastTransitionTime":"2025-11-24T06:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.297011 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.297081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.297099 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.297124 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.297145 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:13Z","lastTransitionTime":"2025-11-24T06:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.401143 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.401238 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.401266 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.401300 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.401323 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:13Z","lastTransitionTime":"2025-11-24T06:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.505312 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.505393 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.505413 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.505451 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.505477 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:13Z","lastTransitionTime":"2025-11-24T06:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.609097 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.609169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.609187 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.609215 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.609235 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:13Z","lastTransitionTime":"2025-11-24T06:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.628071 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.628078 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:13 crc kubenswrapper[4799]: E1124 06:48:13.628247 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.628112 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:13 crc kubenswrapper[4799]: E1124 06:48:13.628370 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:13 crc kubenswrapper[4799]: E1124 06:48:13.628654 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.712709 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.712776 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.712801 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.712830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.712881 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:13Z","lastTransitionTime":"2025-11-24T06:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.816408 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.816489 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.816509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.816542 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.816563 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:13Z","lastTransitionTime":"2025-11-24T06:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.919991 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.920053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.920062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.920079 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:13 crc kubenswrapper[4799]: I1124 06:48:13.920093 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:13Z","lastTransitionTime":"2025-11-24T06:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.028463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.028550 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.028577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.028612 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.028636 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:14Z","lastTransitionTime":"2025-11-24T06:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.130982 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.131334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.131447 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.131628 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.131734 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:14Z","lastTransitionTime":"2025-11-24T06:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.234716 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.234774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.234789 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.234810 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.234823 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:14Z","lastTransitionTime":"2025-11-24T06:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.338069 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.338137 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.338154 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.338178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.338198 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:14Z","lastTransitionTime":"2025-11-24T06:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.441134 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.441178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.441190 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.441209 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.441224 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:14Z","lastTransitionTime":"2025-11-24T06:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.544280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.544664 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.544729 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.544823 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.544964 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:14Z","lastTransitionTime":"2025-11-24T06:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.628005 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:14 crc kubenswrapper[4799]: E1124 06:48:14.628474 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.649108 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.649179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.649204 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.649237 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.649262 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:14Z","lastTransitionTime":"2025-11-24T06:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.753116 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.753161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.753170 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.753186 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.753197 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:14Z","lastTransitionTime":"2025-11-24T06:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.856478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.856540 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.856553 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.856576 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.856590 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:14Z","lastTransitionTime":"2025-11-24T06:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.959080 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.959435 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.959572 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.959677 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:14 crc kubenswrapper[4799]: I1124 06:48:14.959752 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:14Z","lastTransitionTime":"2025-11-24T06:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.030728 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:15 crc kubenswrapper[4799]: E1124 06:48:15.030919 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:15 crc kubenswrapper[4799]: E1124 06:48:15.030992 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs podName:f587ff07-479b-4c95-95cb-406faffcfb68 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:19.030974824 +0000 UTC m=+44.686957298 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs") pod "network-metrics-daemon-ldbr4" (UID: "f587ff07-479b-4c95-95cb-406faffcfb68") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.063196 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.063233 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.063242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.063256 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.063265 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:15Z","lastTransitionTime":"2025-11-24T06:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.166055 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.166431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.166521 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.166591 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.166674 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:15Z","lastTransitionTime":"2025-11-24T06:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.269775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.269875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.269921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.269944 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.269966 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:15Z","lastTransitionTime":"2025-11-24T06:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.372160 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.372212 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.372224 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.372241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.372253 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:15Z","lastTransitionTime":"2025-11-24T06:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.474580 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.474621 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.474631 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.474647 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.474658 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:15Z","lastTransitionTime":"2025-11-24T06:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.577053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.577095 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.577105 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.577120 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.577130 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:15Z","lastTransitionTime":"2025-11-24T06:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.627730 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.627784 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.627772 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:15 crc kubenswrapper[4799]: E1124 06:48:15.627943 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:15 crc kubenswrapper[4799]: E1124 06:48:15.628054 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:15 crc kubenswrapper[4799]: E1124 06:48:15.628144 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.643002 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.656320 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.668057 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.679286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.679333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.679342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.679355 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.679363 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:15Z","lastTransitionTime":"2025-11-24T06:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.681968 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.692972 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.706660 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.722131 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.738243 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.757824 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"173564 6243 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1124 06:48:09.173566 6243 lb_config.go:1031] Cluster endpoints for openshift-dns/dns-default for network=default are: map[]\\\\nI1124 06:48:09.173571 6243 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1124 06:48:09.173202 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:48:09.173585 6243 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-ch\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.771880 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.781521 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.781562 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.781573 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.781591 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.781602 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:15Z","lastTransitionTime":"2025-11-24T06:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.783921 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.796148 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.809655 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.825329 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.845777 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.859362 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.884769 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.884835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.884908 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.884938 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.884957 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:15Z","lastTransitionTime":"2025-11-24T06:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.885114 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:15Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.988677 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.988757 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.988778 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.988805 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:15 crc kubenswrapper[4799]: I1124 06:48:15.988823 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:15Z","lastTransitionTime":"2025-11-24T06:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.132159 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.132262 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.132283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.132312 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.132331 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:16Z","lastTransitionTime":"2025-11-24T06:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.236965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.237201 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.237228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.237259 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.237279 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:16Z","lastTransitionTime":"2025-11-24T06:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.340098 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.340152 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.340163 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.340203 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.340215 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:16Z","lastTransitionTime":"2025-11-24T06:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.443111 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.443157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.443169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.443187 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.443205 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:16Z","lastTransitionTime":"2025-11-24T06:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.546406 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.546458 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.546467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.546484 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.546495 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:16Z","lastTransitionTime":"2025-11-24T06:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.627522 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:16 crc kubenswrapper[4799]: E1124 06:48:16.627718 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.649450 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.649495 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.649504 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.649525 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.649542 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:16Z","lastTransitionTime":"2025-11-24T06:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.768993 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.769046 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.769057 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.769078 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.769089 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:16Z","lastTransitionTime":"2025-11-24T06:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.871466 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.871514 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.871531 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.871550 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.871567 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:16Z","lastTransitionTime":"2025-11-24T06:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.974167 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.974511 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.974531 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.974556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:16 crc kubenswrapper[4799]: I1124 06:48:16.974574 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:16Z","lastTransitionTime":"2025-11-24T06:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.076765 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.076812 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.076825 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.076841 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.076873 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:17Z","lastTransitionTime":"2025-11-24T06:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.180069 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.180221 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.180242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.180268 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.180325 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:17Z","lastTransitionTime":"2025-11-24T06:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.283105 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.283167 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.283191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.283215 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.283239 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:17Z","lastTransitionTime":"2025-11-24T06:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.386529 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.386591 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.386609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.386634 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.386655 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:17Z","lastTransitionTime":"2025-11-24T06:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.490375 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.490461 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.490487 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.490525 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.490556 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:17Z","lastTransitionTime":"2025-11-24T06:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.593750 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.593822 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.593836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.593888 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.593904 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:17Z","lastTransitionTime":"2025-11-24T06:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.628166 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.628208 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.628297 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:17 crc kubenswrapper[4799]: E1124 06:48:17.628349 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:17 crc kubenswrapper[4799]: E1124 06:48:17.628447 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:17 crc kubenswrapper[4799]: E1124 06:48:17.628559 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.697800 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.697905 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.697924 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.697952 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.697970 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:17Z","lastTransitionTime":"2025-11-24T06:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.800946 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.801008 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.801027 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.801056 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.801076 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:17Z","lastTransitionTime":"2025-11-24T06:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.905866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.905950 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.905965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.905986 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:17 crc kubenswrapper[4799]: I1124 06:48:17.906004 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:17Z","lastTransitionTime":"2025-11-24T06:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.008525 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.008589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.008601 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.008615 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.008625 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:18Z","lastTransitionTime":"2025-11-24T06:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.112102 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.112161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.112171 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.112192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.112204 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:18Z","lastTransitionTime":"2025-11-24T06:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.215152 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.215213 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.215224 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.215242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.215254 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:18Z","lastTransitionTime":"2025-11-24T06:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.317656 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.317699 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.317711 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.317728 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.317741 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:18Z","lastTransitionTime":"2025-11-24T06:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.420280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.420320 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.420331 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.420349 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.420361 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:18Z","lastTransitionTime":"2025-11-24T06:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.522306 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.522354 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.522365 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.522381 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.522391 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:18Z","lastTransitionTime":"2025-11-24T06:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.625620 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.625671 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.625688 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.625703 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.625714 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:18Z","lastTransitionTime":"2025-11-24T06:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.628047 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:18 crc kubenswrapper[4799]: E1124 06:48:18.628161 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.728496 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.728554 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.728567 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.728584 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.728594 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:18Z","lastTransitionTime":"2025-11-24T06:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.831235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.831277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.831287 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.831303 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.831315 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:18Z","lastTransitionTime":"2025-11-24T06:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.933531 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.933598 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.933611 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.933629 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:18 crc kubenswrapper[4799]: I1124 06:48:18.933642 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:18Z","lastTransitionTime":"2025-11-24T06:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.037153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.037241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.037268 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.037305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.037327 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:19Z","lastTransitionTime":"2025-11-24T06:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.077485 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:19 crc kubenswrapper[4799]: E1124 06:48:19.077779 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:19 crc kubenswrapper[4799]: E1124 06:48:19.077929 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs podName:f587ff07-479b-4c95-95cb-406faffcfb68 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:27.077897763 +0000 UTC m=+52.733880267 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs") pod "network-metrics-daemon-ldbr4" (UID: "f587ff07-479b-4c95-95cb-406faffcfb68") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.141505 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.141598 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.141624 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.141663 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.141689 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:19Z","lastTransitionTime":"2025-11-24T06:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.245340 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.245411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.245428 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.245452 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.245471 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:19Z","lastTransitionTime":"2025-11-24T06:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.347830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.347900 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.347911 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.347932 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.347944 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:19Z","lastTransitionTime":"2025-11-24T06:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.451420 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.451480 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.451494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.451513 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.451527 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:19Z","lastTransitionTime":"2025-11-24T06:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.554328 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.554402 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.554420 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.554450 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.554472 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:19Z","lastTransitionTime":"2025-11-24T06:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.628245 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.628308 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.628345 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:19 crc kubenswrapper[4799]: E1124 06:48:19.628490 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:19 crc kubenswrapper[4799]: E1124 06:48:19.628602 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:19 crc kubenswrapper[4799]: E1124 06:48:19.628761 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.657580 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.657651 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.657673 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.657700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.657723 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:19Z","lastTransitionTime":"2025-11-24T06:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.761946 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.762018 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.762046 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.762081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.762106 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:19Z","lastTransitionTime":"2025-11-24T06:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.865581 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.865641 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.865657 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.865683 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.865894 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:19Z","lastTransitionTime":"2025-11-24T06:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.969021 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.969065 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.969074 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.969121 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:19 crc kubenswrapper[4799]: I1124 06:48:19.969137 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:19Z","lastTransitionTime":"2025-11-24T06:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.072115 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.072177 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.072196 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.072220 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.072238 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:20Z","lastTransitionTime":"2025-11-24T06:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.175490 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.175548 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.175567 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.175593 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.175612 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:20Z","lastTransitionTime":"2025-11-24T06:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.278367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.278456 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.278476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.278511 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.278532 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:20Z","lastTransitionTime":"2025-11-24T06:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.381386 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.381468 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.381485 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.381512 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.381527 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:20Z","lastTransitionTime":"2025-11-24T06:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.484761 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.485119 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.485208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.485319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.485396 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:20Z","lastTransitionTime":"2025-11-24T06:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.588791 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.589229 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.589295 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.589364 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.589430 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:20Z","lastTransitionTime":"2025-11-24T06:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.627553 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:20 crc kubenswrapper[4799]: E1124 06:48:20.627756 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.693298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.693369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.693394 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.693430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.693457 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:20Z","lastTransitionTime":"2025-11-24T06:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.797040 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.797130 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.797158 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.797195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.797222 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:20Z","lastTransitionTime":"2025-11-24T06:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.901407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.901493 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.901522 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.901556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:20 crc kubenswrapper[4799]: I1124 06:48:20.901580 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:20Z","lastTransitionTime":"2025-11-24T06:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.004829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.004940 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.004960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.004989 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.005011 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:21Z","lastTransitionTime":"2025-11-24T06:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.108901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.108964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.108982 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.109007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.109026 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:21Z","lastTransitionTime":"2025-11-24T06:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.213760 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.214390 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.214635 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.214814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.215007 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:21Z","lastTransitionTime":"2025-11-24T06:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.318481 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.319139 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.319200 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.319225 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.319239 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:21Z","lastTransitionTime":"2025-11-24T06:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.422133 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.422284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.422311 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.422352 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.422377 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:21Z","lastTransitionTime":"2025-11-24T06:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.526494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.526569 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.526597 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.526629 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.526652 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:21Z","lastTransitionTime":"2025-11-24T06:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.627619 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:21 crc kubenswrapper[4799]: E1124 06:48:21.627841 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.628233 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.628274 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:21 crc kubenswrapper[4799]: E1124 06:48:21.628387 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:21 crc kubenswrapper[4799]: E1124 06:48:21.628654 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.630497 4799 scope.go:117] "RemoveContainer" containerID="eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.632125 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.632169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.632192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.632221 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.632240 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:21Z","lastTransitionTime":"2025-11-24T06:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.735712 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.735782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.735803 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.735832 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.735881 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:21Z","lastTransitionTime":"2025-11-24T06:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.839840 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.839942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.839961 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.839987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.840010 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:21Z","lastTransitionTime":"2025-11-24T06:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.942894 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.942941 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.942954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.942973 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:21 crc kubenswrapper[4799]: I1124 06:48:21.942986 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:21Z","lastTransitionTime":"2025-11-24T06:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.005591 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/1.log" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.009494 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0"} Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.010300 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.034829 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.045548 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.045587 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.045595 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.045611 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.045623 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.061651 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.080091 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.106054 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.130528 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.149239 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.149291 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.149303 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.149322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.149333 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.151131 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.171599 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.191052 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.205806 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.223871 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"173564 6243 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1124 06:48:09.173566 6243 lb_config.go:1031] Cluster endpoints for openshift-dns/dns-default for network=default are: map[]\\\\nI1124 06:48:09.173571 6243 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1124 06:48:09.173202 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:48:09.173585 6243 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-ch\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.238693 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.252528 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.252588 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.252600 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.252620 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.252633 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.254593 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.268962 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.283791 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.300739 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.316303 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.332666 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.355007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.355062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.355075 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.355095 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.355108 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.458935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.459009 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.459033 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.459063 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.459084 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.546149 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.546216 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.546235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.546262 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.546280 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: E1124 06:48:22.571151 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.577528 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.577590 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.577609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.577636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.577661 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: E1124 06:48:22.594460 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.601001 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.601081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.601108 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.601142 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.601162 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: E1124 06:48:22.620113 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.625981 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.626054 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.626077 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.626112 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.626136 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.627166 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:22 crc kubenswrapper[4799]: E1124 06:48:22.627299 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:22 crc kubenswrapper[4799]: E1124 06:48:22.646056 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.653487 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.653526 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.653538 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.653564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.653578 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: E1124 06:48:22.674167 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:22Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:22 crc kubenswrapper[4799]: E1124 06:48:22.674415 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.677364 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.677423 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.677442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.677472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.677489 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.781397 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.781476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.781497 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.781528 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.781551 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.884380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.884445 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.884463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.884490 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.884508 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.987337 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.987411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.987435 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.987465 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:22 crc kubenswrapper[4799]: I1124 06:48:22.987486 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:22Z","lastTransitionTime":"2025-11-24T06:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.016980 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/2.log" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.018369 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/1.log" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.022367 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0" exitCode=1 Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.022438 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0"} Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.022513 4799 scope.go:117] "RemoveContainer" containerID="eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.023467 4799 scope.go:117] "RemoveContainer" containerID="dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0" Nov 24 06:48:23 crc kubenswrapper[4799]: E1124 06:48:23.023772 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.045071 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.059767 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.082240 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.090709 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.090783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.090797 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.090818 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.090833 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:23Z","lastTransitionTime":"2025-11-24T06:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.104676 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.123308 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.144309 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.164081 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.194640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.194707 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.194726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.194757 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.194778 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:23Z","lastTransitionTime":"2025-11-24T06:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.201553 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.222996 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.248562 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.271073 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.298060 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.299327 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.299401 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.299415 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.299439 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.299454 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:23Z","lastTransitionTime":"2025-11-24T06:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.317380 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.341933 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.366031 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.390051 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.403386 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.403491 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.403510 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.403537 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.403585 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:23Z","lastTransitionTime":"2025-11-24T06:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.425037 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eafed5c9cd84ae24b360c423863b8d666caab387de38c1777de5a8aa5548f7fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"message\\\":\\\"173564 6243 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1124 06:48:09.173566 6243 lb_config.go:1031] Cluster endpoints for openshift-dns/dns-default for network=default are: map[]\\\\nI1124 06:48:09.173571 6243 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1124 06:48:09.173202 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:09Z is after 2025-08-24T17:21:41Z]\\\\nI1124 06:48:09.173585 6243 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-ch\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"t handler 2 for removal\\\\nI1124 06:48:22.681331 6439 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:22.681337 6439 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681376 6439 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 06:48:22.681251 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681402 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 06:48:22.681399 6439 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681604 6439 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:48:22.681615 6439 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681776 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681629 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:22.682681 6439 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:23Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.507629 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.507762 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.507782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.507816 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.507892 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:23Z","lastTransitionTime":"2025-11-24T06:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.611410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.611476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.611499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.611526 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.611546 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:23Z","lastTransitionTime":"2025-11-24T06:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.628178 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.628286 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.628193 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:23 crc kubenswrapper[4799]: E1124 06:48:23.628442 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:23 crc kubenswrapper[4799]: E1124 06:48:23.628639 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:23 crc kubenswrapper[4799]: E1124 06:48:23.628787 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.715193 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.715275 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.715302 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.715336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.715357 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:23Z","lastTransitionTime":"2025-11-24T06:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.819906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.819990 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.820010 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.820041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.820062 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:23Z","lastTransitionTime":"2025-11-24T06:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.923484 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.923543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.923556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.923580 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:23 crc kubenswrapper[4799]: I1124 06:48:23.923593 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:23Z","lastTransitionTime":"2025-11-24T06:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.026332 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.026411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.026424 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.026441 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.026476 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:24Z","lastTransitionTime":"2025-11-24T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.028236 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/2.log" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.033248 4799 scope.go:117] "RemoveContainer" containerID="dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0" Nov 24 06:48:24 crc kubenswrapper[4799]: E1124 06:48:24.033458 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.049468 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.073988 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.100683 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.128718 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.130953 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.131017 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.131032 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.131061 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.131077 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:24Z","lastTransitionTime":"2025-11-24T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.156305 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.179565 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.201777 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.226537 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.235540 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.235646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.235661 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.235686 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.235703 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:24Z","lastTransitionTime":"2025-11-24T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.252378 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"t handler 2 for removal\\\\nI1124 06:48:22.681331 6439 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:22.681337 6439 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681376 6439 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 06:48:22.681251 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681402 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 06:48:22.681399 6439 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681604 6439 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:48:22.681615 6439 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681776 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681629 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:22.682681 6439 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.268768 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.286195 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.304227 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.316562 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.333615 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.339448 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.339490 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.339499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.339515 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.339523 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:24Z","lastTransitionTime":"2025-11-24T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.352266 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.366068 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.390255 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:24Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.443176 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.443256 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.443298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.443317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.443328 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:24Z","lastTransitionTime":"2025-11-24T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.547214 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.547335 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.547354 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.547418 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.547437 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:24Z","lastTransitionTime":"2025-11-24T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.627719 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:24 crc kubenswrapper[4799]: E1124 06:48:24.628080 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.651382 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.651453 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.651463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.651480 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.651491 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:24Z","lastTransitionTime":"2025-11-24T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.755179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.755241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.755258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.755292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.755312 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:24Z","lastTransitionTime":"2025-11-24T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.859292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.859404 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.859422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.859446 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.859464 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:24Z","lastTransitionTime":"2025-11-24T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.962809 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.962870 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.962880 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.962898 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:24 crc kubenswrapper[4799]: I1124 06:48:24.962909 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:24Z","lastTransitionTime":"2025-11-24T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.065739 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.065816 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.065829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.065895 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.065911 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:25Z","lastTransitionTime":"2025-11-24T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.169638 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.169752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.169773 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.169829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.169895 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:25Z","lastTransitionTime":"2025-11-24T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.273743 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.273817 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.273829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.273883 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.273899 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:25Z","lastTransitionTime":"2025-11-24T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.377646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.377722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.377745 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.377782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.377809 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:25Z","lastTransitionTime":"2025-11-24T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.480782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.480838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.480874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.480895 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.480908 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:25Z","lastTransitionTime":"2025-11-24T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.584376 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.584441 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.584453 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.584473 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.584487 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:25Z","lastTransitionTime":"2025-11-24T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.627676 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.627757 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.627876 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:25 crc kubenswrapper[4799]: E1124 06:48:25.627897 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:25 crc kubenswrapper[4799]: E1124 06:48:25.627972 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:25 crc kubenswrapper[4799]: E1124 06:48:25.628041 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.646519 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.662430 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.681696 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.689606 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.689653 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.689665 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.689684 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.689702 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:25Z","lastTransitionTime":"2025-11-24T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.698154 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.714705 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.736227 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.753384 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.781928 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.792204 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.792247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.792258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.792276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.792288 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:25Z","lastTransitionTime":"2025-11-24T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.801250 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.817686 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.828775 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.853408 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.869061 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.885064 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.895520 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.895573 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.895596 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.895635 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.895663 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:25Z","lastTransitionTime":"2025-11-24T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.901715 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.915580 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.940400 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"t handler 2 for removal\\\\nI1124 06:48:22.681331 6439 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:22.681337 6439 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681376 6439 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 06:48:22.681251 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681402 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 06:48:22.681399 6439 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681604 6439 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:48:22.681615 6439 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681776 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681629 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:22.682681 6439 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:25Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.999319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.999385 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.999404 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.999432 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:25 crc kubenswrapper[4799]: I1124 06:48:25.999462 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:25Z","lastTransitionTime":"2025-11-24T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.102985 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.103034 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.103046 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.103063 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.103076 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:26Z","lastTransitionTime":"2025-11-24T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.207165 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.207240 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.207258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.207286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.207308 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:26Z","lastTransitionTime":"2025-11-24T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.310197 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.310235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.310248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.310273 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.310297 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:26Z","lastTransitionTime":"2025-11-24T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.412783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.412821 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.412831 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.412863 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.412877 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:26Z","lastTransitionTime":"2025-11-24T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.516382 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.516444 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.516461 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.516490 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.516507 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:26Z","lastTransitionTime":"2025-11-24T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.619118 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.619171 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.619184 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.619203 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.619217 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:26Z","lastTransitionTime":"2025-11-24T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.627693 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:26 crc kubenswrapper[4799]: E1124 06:48:26.627826 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.721766 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.721808 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.721820 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.721837 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.721877 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:26Z","lastTransitionTime":"2025-11-24T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.825184 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.825245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.825343 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.825380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.825399 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:26Z","lastTransitionTime":"2025-11-24T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.928932 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.928974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.928983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.928996 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:26 crc kubenswrapper[4799]: I1124 06:48:26.929007 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:26Z","lastTransitionTime":"2025-11-24T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.033317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.033374 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.033386 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.033407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.033420 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:27Z","lastTransitionTime":"2025-11-24T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.110905 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.111272 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.111440 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs podName:f587ff07-479b-4c95-95cb-406faffcfb68 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:43.111397821 +0000 UTC m=+68.767380325 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs") pod "network-metrics-daemon-ldbr4" (UID: "f587ff07-479b-4c95-95cb-406faffcfb68") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.137347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.137451 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.137534 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.137567 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.137592 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:27Z","lastTransitionTime":"2025-11-24T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.240655 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.240736 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.240759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.240791 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.240814 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:27Z","lastTransitionTime":"2025-11-24T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.344465 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.344529 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.344549 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.344576 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.344596 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:27Z","lastTransitionTime":"2025-11-24T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.414541 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.414870 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:48:59.414793134 +0000 UTC m=+85.070775648 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.447964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.448036 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.448057 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.448084 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.448111 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:27Z","lastTransitionTime":"2025-11-24T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.516015 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.516080 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.516121 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.516175 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516330 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516373 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516379 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516441 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:59.516411955 +0000 UTC m=+85.172394459 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516453 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516575 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516608 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516402 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516704 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516533 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:59.516495067 +0000 UTC m=+85.172477711 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516825 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:59.516808976 +0000 UTC m=+85.172791490 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.516883 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:48:59.516836677 +0000 UTC m=+85.172819191 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.550936 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.550986 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.551008 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.551036 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.551057 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:27Z","lastTransitionTime":"2025-11-24T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.629168 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.629261 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.629422 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.629530 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.629757 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:27 crc kubenswrapper[4799]: E1124 06:48:27.629924 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.654889 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.654945 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.654968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.654993 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.655015 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:27Z","lastTransitionTime":"2025-11-24T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.758077 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.758166 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.758192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.758228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.758254 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:27Z","lastTransitionTime":"2025-11-24T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.861508 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.861553 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.861564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.861579 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.861590 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:27Z","lastTransitionTime":"2025-11-24T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.965650 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.965718 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.965742 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.965771 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:27 crc kubenswrapper[4799]: I1124 06:48:27.965795 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:27Z","lastTransitionTime":"2025-11-24T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.069731 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.069818 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.069833 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.069916 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.069933 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:28Z","lastTransitionTime":"2025-11-24T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.172962 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.173527 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.173556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.173594 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.173620 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:28Z","lastTransitionTime":"2025-11-24T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.277686 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.277775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.277803 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.277840 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.277911 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:28Z","lastTransitionTime":"2025-11-24T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.381016 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.381089 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.381105 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.381131 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.381152 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:28Z","lastTransitionTime":"2025-11-24T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.485060 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.485132 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.485153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.485181 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.485200 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:28Z","lastTransitionTime":"2025-11-24T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.589039 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.589132 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.589151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.589180 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.589206 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:28Z","lastTransitionTime":"2025-11-24T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.627946 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:28 crc kubenswrapper[4799]: E1124 06:48:28.628171 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.693549 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.693723 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.694219 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.694279 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.694310 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:28Z","lastTransitionTime":"2025-11-24T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.798579 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.798630 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.798640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.798654 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.798665 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:28Z","lastTransitionTime":"2025-11-24T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.902394 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.902452 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.902464 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.902479 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:28 crc kubenswrapper[4799]: I1124 06:48:28.902490 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:28Z","lastTransitionTime":"2025-11-24T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.006003 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.006083 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.006103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.006129 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.006150 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:29Z","lastTransitionTime":"2025-11-24T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.109780 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.109835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.109904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.109934 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.109955 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:29Z","lastTransitionTime":"2025-11-24T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.213164 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.213234 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.213251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.213276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.213301 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:29Z","lastTransitionTime":"2025-11-24T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.317336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.317411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.317432 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.317461 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.317482 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:29Z","lastTransitionTime":"2025-11-24T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.421704 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.421792 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.421811 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.421905 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.421931 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:29Z","lastTransitionTime":"2025-11-24T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.524945 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.525009 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.525028 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.525055 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.525075 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:29Z","lastTransitionTime":"2025-11-24T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.627445 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:29 crc kubenswrapper[4799]: E1124 06:48:29.627696 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.627456 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.628083 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.628141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.628164 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.628187 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.628239 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:29Z","lastTransitionTime":"2025-11-24T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.627910 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:29 crc kubenswrapper[4799]: E1124 06:48:29.628618 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:29 crc kubenswrapper[4799]: E1124 06:48:29.628749 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.649167 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.665176 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.678680 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.695694 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.715156 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.741239 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.741279 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.741290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.741305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.741315 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:29Z","lastTransitionTime":"2025-11-24T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.761520 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.805451 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.820164 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.829653 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.844178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.844267 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.844302 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.844319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.844331 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:29Z","lastTransitionTime":"2025-11-24T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.846954 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.860198 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.881912 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"t handler 2 for removal\\\\nI1124 06:48:22.681331 6439 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:22.681337 6439 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681376 6439 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 06:48:22.681251 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681402 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 06:48:22.681399 6439 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681604 6439 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:48:22.681615 6439 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681776 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681629 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:22.682681 6439 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.899256 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.914136 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.931475 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.945053 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.947815 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.947904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.947923 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.947951 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.947970 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:29Z","lastTransitionTime":"2025-11-24T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.958172 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.975950 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:29 crc kubenswrapper[4799]: I1124 06:48:29.993063 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:29Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.051746 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.051803 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.051819 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.051873 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.051892 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:30Z","lastTransitionTime":"2025-11-24T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.155276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.155340 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.155354 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.155378 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.155392 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:30Z","lastTransitionTime":"2025-11-24T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.258878 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.258949 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.258973 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.259007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.259033 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:30Z","lastTransitionTime":"2025-11-24T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.362402 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.362473 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.362491 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.362517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.362535 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:30Z","lastTransitionTime":"2025-11-24T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.465395 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.465492 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.465518 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.465556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.465583 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:30Z","lastTransitionTime":"2025-11-24T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.569066 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.569263 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.569326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.569356 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.569376 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:30Z","lastTransitionTime":"2025-11-24T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.628040 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:30 crc kubenswrapper[4799]: E1124 06:48:30.628318 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.673381 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.673445 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.673465 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.673493 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.673513 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:30Z","lastTransitionTime":"2025-11-24T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.777915 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.778003 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.778027 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.778057 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.778079 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:30Z","lastTransitionTime":"2025-11-24T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.881523 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.881650 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.881672 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.881705 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.881726 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:30Z","lastTransitionTime":"2025-11-24T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.985053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.985112 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.985131 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.985156 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:30 crc kubenswrapper[4799]: I1124 06:48:30.985176 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:30Z","lastTransitionTime":"2025-11-24T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.088724 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.088795 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.088814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.088841 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.088884 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:31Z","lastTransitionTime":"2025-11-24T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.192096 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.192172 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.192191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.192221 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.192244 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:31Z","lastTransitionTime":"2025-11-24T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.295026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.295083 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.295093 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.295111 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.295121 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:31Z","lastTransitionTime":"2025-11-24T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.397919 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.397986 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.397999 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.398013 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.398023 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:31Z","lastTransitionTime":"2025-11-24T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.502577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.502767 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.502787 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.502815 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.502837 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:31Z","lastTransitionTime":"2025-11-24T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.606296 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.606401 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.606427 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.606464 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.606493 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:31Z","lastTransitionTime":"2025-11-24T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.627920 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.628028 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.628092 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:31 crc kubenswrapper[4799]: E1124 06:48:31.628358 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:31 crc kubenswrapper[4799]: E1124 06:48:31.628547 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:31 crc kubenswrapper[4799]: E1124 06:48:31.628772 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.711146 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.711232 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.711283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.711320 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.711346 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:31Z","lastTransitionTime":"2025-11-24T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.815006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.815098 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.815121 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.815150 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.815172 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:31Z","lastTransitionTime":"2025-11-24T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.920515 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.920684 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.920712 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.920779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:31 crc kubenswrapper[4799]: I1124 06:48:31.920801 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:31Z","lastTransitionTime":"2025-11-24T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.023721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.023786 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.023803 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.023829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.023877 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.127286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.127358 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.127378 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.127406 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.127426 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.230959 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.231023 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.231041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.231068 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.231086 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.334113 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.334192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.334210 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.334239 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.334261 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.437463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.437547 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.437569 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.437600 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.437622 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.541543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.541636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.541656 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.541685 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.541706 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.627564 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:32 crc kubenswrapper[4799]: E1124 06:48:32.627815 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.646023 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.646106 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.646155 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.646235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.646254 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.750838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.750973 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.750999 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.751516 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.751724 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.780203 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.780337 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.780368 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.780409 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.780435 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: E1124 06:48:32.803026 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.809787 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.809870 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.809902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.809950 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.809976 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: E1124 06:48:32.831696 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.838305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.838359 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.838407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.838435 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.838524 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: E1124 06:48:32.860202 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.865630 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.865695 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.865708 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.865730 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.865743 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: E1124 06:48:32.883175 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.887585 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.887624 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.887640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.887660 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.887675 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:32 crc kubenswrapper[4799]: E1124 06:48:32.907279 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:32Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:32 crc kubenswrapper[4799]: E1124 06:48:32.907442 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.910287 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.910347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.910365 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.910391 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:32 crc kubenswrapper[4799]: I1124 06:48:32.910410 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:32Z","lastTransitionTime":"2025-11-24T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.013789 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.013892 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.013921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.013953 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.013979 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:33Z","lastTransitionTime":"2025-11-24T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.118294 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.118353 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.118368 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.118391 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.118410 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:33Z","lastTransitionTime":"2025-11-24T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.234046 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.234092 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.234106 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.234127 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.234144 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:33Z","lastTransitionTime":"2025-11-24T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.336400 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.336446 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.336454 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.336467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.336478 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:33Z","lastTransitionTime":"2025-11-24T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.439175 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.439230 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.439242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.439262 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.439284 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:33Z","lastTransitionTime":"2025-11-24T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.543213 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.543262 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.543277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.543298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.543314 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:33Z","lastTransitionTime":"2025-11-24T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.628128 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.628267 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:33 crc kubenswrapper[4799]: E1124 06:48:33.628274 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.628125 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:33 crc kubenswrapper[4799]: E1124 06:48:33.628819 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:33 crc kubenswrapper[4799]: E1124 06:48:33.628557 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.646642 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.646710 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.646727 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.646752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.646770 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:33Z","lastTransitionTime":"2025-11-24T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.750955 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.751041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.751062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.751092 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.751115 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:33Z","lastTransitionTime":"2025-11-24T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.855295 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.855367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.855384 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.855411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.855431 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:33Z","lastTransitionTime":"2025-11-24T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.957668 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.957708 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.957717 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.957730 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:33 crc kubenswrapper[4799]: I1124 06:48:33.957739 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:33Z","lastTransitionTime":"2025-11-24T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:34 crc kubenswrapper[4799]: I1124 06:48:34.061531 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:34 crc kubenswrapper[4799]: I1124 06:48:34.061597 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:34 crc kubenswrapper[4799]: I1124 06:48:34.061621 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:34 crc kubenswrapper[4799]: I1124 06:48:34.061652 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:34 crc kubenswrapper[4799]: I1124 06:48:34.061677 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:34Z","lastTransitionTime":"2025-11-24T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.168074 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.168166 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:35 crc kubenswrapper[4799]: E1124 06:48:35.168279 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.168076 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.168379 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:35 crc kubenswrapper[4799]: E1124 06:48:35.168498 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:35 crc kubenswrapper[4799]: E1124 06:48:35.168614 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:35 crc kubenswrapper[4799]: E1124 06:48:35.168777 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.170964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.171008 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.171025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.171044 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.171058 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:35Z","lastTransitionTime":"2025-11-24T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.275867 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.276000 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.276023 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.276052 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.276072 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:35Z","lastTransitionTime":"2025-11-24T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.379741 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.379806 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.379821 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.379840 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.379870 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:35Z","lastTransitionTime":"2025-11-24T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.482974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.483061 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.483073 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.483096 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.483108 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:35Z","lastTransitionTime":"2025-11-24T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.585604 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.585658 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.585674 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.585698 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.585715 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:35Z","lastTransitionTime":"2025-11-24T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.654573 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ab620d5-4762-43c6-b1b7-906e670217ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b3642617df8fcafa59d2eb42352cfe1f5a455e0a702843c2fffc8491cde5c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b888a1477657d057e486b872a2a3da23225070bfee8b34139b33f4762c4d32fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e69e47ea9ea59fe65c19fad4217224390526bb50c23984de97db2cefa6cfe0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.689881 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.689950 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.689965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.689987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.690006 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:35Z","lastTransitionTime":"2025-11-24T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.692685 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.712977 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.737269 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.754490 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.778652 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.793747 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.793814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.793836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.793902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.793930 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:35Z","lastTransitionTime":"2025-11-24T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.801519 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.822613 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.834460 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.851322 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.882098 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"t handler 2 for removal\\\\nI1124 06:48:22.681331 6439 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:22.681337 6439 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681376 6439 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 06:48:22.681251 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681402 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 06:48:22.681399 6439 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681604 6439 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:48:22.681615 6439 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681776 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681629 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:22.682681 6439 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.897015 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.897082 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.897101 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.897131 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.897151 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:35Z","lastTransitionTime":"2025-11-24T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.904218 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.926304 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.943797 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.959014 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.973111 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:35 crc kubenswrapper[4799]: I1124 06:48:35.993908 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:35Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.002586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.002637 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.002652 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.002675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.002690 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:36Z","lastTransitionTime":"2025-11-24T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.010080 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:36Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.105960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.106008 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.106017 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.106035 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.106047 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:36Z","lastTransitionTime":"2025-11-24T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.209022 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.209120 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.209168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.209208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.209234 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:36Z","lastTransitionTime":"2025-11-24T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.313102 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.313173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.313192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.313220 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.313238 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:36Z","lastTransitionTime":"2025-11-24T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.416431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.416539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.416559 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.416586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.416605 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:36Z","lastTransitionTime":"2025-11-24T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.520174 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.520242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.520253 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.520270 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.520284 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:36Z","lastTransitionTime":"2025-11-24T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.623914 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.623956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.623969 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.623987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.624001 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:36Z","lastTransitionTime":"2025-11-24T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.627510 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.627576 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:36 crc kubenswrapper[4799]: E1124 06:48:36.627664 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.627585 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:36 crc kubenswrapper[4799]: E1124 06:48:36.627787 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:36 crc kubenswrapper[4799]: E1124 06:48:36.627985 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.628587 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:36 crc kubenswrapper[4799]: E1124 06:48:36.628892 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.727484 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.728040 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.728052 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.728071 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.728083 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:36Z","lastTransitionTime":"2025-11-24T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.831552 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.831635 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.831662 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.831698 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.831725 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:36Z","lastTransitionTime":"2025-11-24T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.934758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.934897 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.934937 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.934974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:36 crc kubenswrapper[4799]: I1124 06:48:36.934998 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:36Z","lastTransitionTime":"2025-11-24T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.039537 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.039625 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.039644 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.039671 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.039691 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:37Z","lastTransitionTime":"2025-11-24T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.143802 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.143904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.143929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.143965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.143987 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:37Z","lastTransitionTime":"2025-11-24T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.247370 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.247444 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.247463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.247493 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.247514 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:37Z","lastTransitionTime":"2025-11-24T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.351385 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.351485 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.351504 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.351544 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.351572 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:37Z","lastTransitionTime":"2025-11-24T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.454667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.454731 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.454752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.454777 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.454797 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:37Z","lastTransitionTime":"2025-11-24T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.558273 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.558363 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.558387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.558424 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.558451 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:37Z","lastTransitionTime":"2025-11-24T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.662226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.662307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.662327 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.662364 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.662384 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:37Z","lastTransitionTime":"2025-11-24T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.765447 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.765532 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.765545 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.765568 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.765597 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:37Z","lastTransitionTime":"2025-11-24T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.868652 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.868711 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.868722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.868741 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.868754 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:37Z","lastTransitionTime":"2025-11-24T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.972994 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.973081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.973107 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.973143 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:37 crc kubenswrapper[4799]: I1124 06:48:37.973169 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:37Z","lastTransitionTime":"2025-11-24T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.077721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.077785 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.077804 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.077833 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.077885 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:38Z","lastTransitionTime":"2025-11-24T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.180750 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.180805 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.180818 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.180838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.180872 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:38Z","lastTransitionTime":"2025-11-24T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.283713 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.283802 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.283829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.283905 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.283936 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:38Z","lastTransitionTime":"2025-11-24T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.386675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.386778 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.386796 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.386822 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.386840 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:38Z","lastTransitionTime":"2025-11-24T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.490308 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.490393 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.490418 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.490452 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.490475 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:38Z","lastTransitionTime":"2025-11-24T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.594543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.594617 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.594642 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.594674 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.594698 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:38Z","lastTransitionTime":"2025-11-24T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.628030 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:38 crc kubenswrapper[4799]: E1124 06:48:38.628238 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.628507 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:38 crc kubenswrapper[4799]: E1124 06:48:38.628613 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.628923 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:38 crc kubenswrapper[4799]: E1124 06:48:38.629038 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.629267 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:38 crc kubenswrapper[4799]: E1124 06:48:38.629380 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.698193 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.698258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.698276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.698303 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.698324 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:38Z","lastTransitionTime":"2025-11-24T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.801663 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.801726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.801745 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.801774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.801792 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:38Z","lastTransitionTime":"2025-11-24T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.906904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.907003 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.907028 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.907062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:38 crc kubenswrapper[4799]: I1124 06:48:38.907084 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:38Z","lastTransitionTime":"2025-11-24T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.010370 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.010463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.010491 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.010529 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.010554 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:39Z","lastTransitionTime":"2025-11-24T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.114297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.114369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.114386 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.114413 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.114435 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:39Z","lastTransitionTime":"2025-11-24T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.217429 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.217511 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.217531 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.217565 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.217592 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:39Z","lastTransitionTime":"2025-11-24T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.320318 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.320352 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.320361 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.320375 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.320385 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:39Z","lastTransitionTime":"2025-11-24T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.423936 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.424002 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.424012 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.424030 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.424039 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:39Z","lastTransitionTime":"2025-11-24T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.527478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.527545 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.527565 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.527592 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.527612 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:39Z","lastTransitionTime":"2025-11-24T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.629901 4799 scope.go:117] "RemoveContainer" containerID="dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0" Nov 24 06:48:39 crc kubenswrapper[4799]: E1124 06:48:39.630205 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.630782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.630827 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.630838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.630872 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.630883 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:39Z","lastTransitionTime":"2025-11-24T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.733182 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.733226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.733235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.733248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.733259 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:39Z","lastTransitionTime":"2025-11-24T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.837433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.837486 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.837500 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.837522 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.837536 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:39Z","lastTransitionTime":"2025-11-24T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.940649 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.940689 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.940702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.940723 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:39 crc kubenswrapper[4799]: I1124 06:48:39.940737 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:39Z","lastTransitionTime":"2025-11-24T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.043421 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.043471 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.043481 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.043501 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.043512 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:40Z","lastTransitionTime":"2025-11-24T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.146654 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.146695 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.146704 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.146722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.146733 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:40Z","lastTransitionTime":"2025-11-24T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.250074 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.250135 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.250153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.250181 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.250203 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:40Z","lastTransitionTime":"2025-11-24T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.353121 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.353173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.353197 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.353228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.353251 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:40Z","lastTransitionTime":"2025-11-24T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.456168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.456235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.456248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.456271 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.456286 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:40Z","lastTransitionTime":"2025-11-24T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.561051 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.561144 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.561166 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.561201 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.561225 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:40Z","lastTransitionTime":"2025-11-24T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.627432 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.627579 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:40 crc kubenswrapper[4799]: E1124 06:48:40.627666 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.627432 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.627580 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:40 crc kubenswrapper[4799]: E1124 06:48:40.627839 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:40 crc kubenswrapper[4799]: E1124 06:48:40.628006 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:40 crc kubenswrapper[4799]: E1124 06:48:40.628142 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.664729 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.664777 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.664788 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.664806 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.664819 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:40Z","lastTransitionTime":"2025-11-24T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.768133 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.768182 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.768196 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.768218 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.768229 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:40Z","lastTransitionTime":"2025-11-24T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.870149 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.870193 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.870208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.870228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.870240 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:40Z","lastTransitionTime":"2025-11-24T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.972775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.972822 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.972831 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.972866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:40 crc kubenswrapper[4799]: I1124 06:48:40.972878 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:40Z","lastTransitionTime":"2025-11-24T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.076800 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.076872 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.076885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.076906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.076919 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:41Z","lastTransitionTime":"2025-11-24T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.179829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.179885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.179896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.179911 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.179923 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:41Z","lastTransitionTime":"2025-11-24T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.283120 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.283179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.283195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.283221 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.283236 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:41Z","lastTransitionTime":"2025-11-24T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.386247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.386313 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.386323 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.386342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.386390 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:41Z","lastTransitionTime":"2025-11-24T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.489064 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.489124 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.489141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.489161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.489172 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:41Z","lastTransitionTime":"2025-11-24T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.591818 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.591887 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.591900 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.591918 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.591936 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:41Z","lastTransitionTime":"2025-11-24T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.695242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.695292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.695307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.695322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.695334 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:41Z","lastTransitionTime":"2025-11-24T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.798410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.798461 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.798474 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.798490 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.798504 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:41Z","lastTransitionTime":"2025-11-24T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.902096 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.902157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.902179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.902199 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:41 crc kubenswrapper[4799]: I1124 06:48:41.902209 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:41Z","lastTransitionTime":"2025-11-24T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.005367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.005412 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.005424 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.005438 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.005450 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:42Z","lastTransitionTime":"2025-11-24T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.108504 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.108559 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.108571 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.108595 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.108609 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:42Z","lastTransitionTime":"2025-11-24T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.210797 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.210872 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.210886 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.210907 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.210924 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:42Z","lastTransitionTime":"2025-11-24T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.313475 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.313512 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.313520 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.313533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.313546 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:42Z","lastTransitionTime":"2025-11-24T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.415961 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.416072 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.416112 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.416129 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.416141 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:42Z","lastTransitionTime":"2025-11-24T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.519009 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.519039 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.519049 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.519062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.519071 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:42Z","lastTransitionTime":"2025-11-24T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.621816 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.621899 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.621912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.621931 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.621948 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:42Z","lastTransitionTime":"2025-11-24T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.628228 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.628310 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:42 crc kubenswrapper[4799]: E1124 06:48:42.628422 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.628325 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:42 crc kubenswrapper[4799]: E1124 06:48:42.628541 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.628336 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:42 crc kubenswrapper[4799]: E1124 06:48:42.628606 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:42 crc kubenswrapper[4799]: E1124 06:48:42.628744 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.724983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.725056 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.725073 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.725095 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.725109 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:42Z","lastTransitionTime":"2025-11-24T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.828589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.828656 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.828672 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.828693 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.828706 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:42Z","lastTransitionTime":"2025-11-24T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.931737 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.931787 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.931799 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.931817 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:42 crc kubenswrapper[4799]: I1124 06:48:42.931829 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:42Z","lastTransitionTime":"2025-11-24T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.001189 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.001240 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.001254 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.001274 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.001289 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: E1124 06:48:43.017923 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:43Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.022367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.022426 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.022437 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.022458 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.022470 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: E1124 06:48:43.038407 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:43Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.043112 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.043156 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.043168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.043194 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.043208 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: E1124 06:48:43.069811 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:43Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.074808 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.074940 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.074958 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.074980 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.074994 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: E1124 06:48:43.088463 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:43Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.093910 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.093955 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.093968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.093987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.094000 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: E1124 06:48:43.110465 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:43Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:43 crc kubenswrapper[4799]: E1124 06:48:43.110629 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.112657 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.112700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.112714 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.112739 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.112752 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.172803 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:43 crc kubenswrapper[4799]: E1124 06:48:43.173198 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:43 crc kubenswrapper[4799]: E1124 06:48:43.173356 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs podName:f587ff07-479b-4c95-95cb-406faffcfb68 nodeName:}" failed. No retries permitted until 2025-11-24 06:49:15.173317068 +0000 UTC m=+100.829299572 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs") pod "network-metrics-daemon-ldbr4" (UID: "f587ff07-479b-4c95-95cb-406faffcfb68") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.215517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.215550 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.215565 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.215586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.215601 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.317609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.317654 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.317668 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.317687 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.317700 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.421512 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.421571 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.421590 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.421616 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.421636 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.524397 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.524440 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.524451 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.524468 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.524480 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.628369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.628411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.628423 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.628439 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.628453 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.731693 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.731772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.731794 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.731824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.731879 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.834413 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.834472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.834483 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.834503 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.834515 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.939050 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.939127 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.939147 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.939179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:43 crc kubenswrapper[4799]: I1124 06:48:43.939201 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:43Z","lastTransitionTime":"2025-11-24T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.042698 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.042758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.042771 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.042790 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.042806 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:44Z","lastTransitionTime":"2025-11-24T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.146177 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.146253 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.146279 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.146317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.146346 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:44Z","lastTransitionTime":"2025-11-24T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.249879 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.249942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.249964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.250003 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.250024 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:44Z","lastTransitionTime":"2025-11-24T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.352566 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.352608 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.352619 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.352634 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.352648 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:44Z","lastTransitionTime":"2025-11-24T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.456184 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.456227 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.456239 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.456261 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.456274 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:44Z","lastTransitionTime":"2025-11-24T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.559744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.559818 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.559835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.559885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.559904 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:44Z","lastTransitionTime":"2025-11-24T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.627731 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.627826 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:44 crc kubenswrapper[4799]: E1124 06:48:44.628035 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.628073 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:44 crc kubenswrapper[4799]: E1124 06:48:44.628266 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.628395 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:44 crc kubenswrapper[4799]: E1124 06:48:44.628427 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:44 crc kubenswrapper[4799]: E1124 06:48:44.628746 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.663221 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.663431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.663521 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.663610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.663700 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:44Z","lastTransitionTime":"2025-11-24T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.766656 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.766728 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.766753 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.766790 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.766815 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:44Z","lastTransitionTime":"2025-11-24T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.869376 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.869425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.869442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.869462 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.869473 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:44Z","lastTransitionTime":"2025-11-24T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.973494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.973550 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.973562 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.973586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:44 crc kubenswrapper[4799]: I1124 06:48:44.973599 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:44Z","lastTransitionTime":"2025-11-24T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.077151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.077209 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.077225 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.077245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.077259 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:45Z","lastTransitionTime":"2025-11-24T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.179723 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.179798 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.179823 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.179886 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.179949 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:45Z","lastTransitionTime":"2025-11-24T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.215349 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvd7f_7980021c-dde3-4c14-a7b6-fbcc947a183d/kube-multus/0.log" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.215405 4799 generic.go:334] "Generic (PLEG): container finished" podID="7980021c-dde3-4c14-a7b6-fbcc947a183d" containerID="ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d" exitCode=1 Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.215443 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvd7f" event={"ID":"7980021c-dde3-4c14-a7b6-fbcc947a183d","Type":"ContainerDied","Data":"ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d"} Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.215885 4799 scope.go:117] "RemoveContainer" containerID="ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.232605 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ab620d5-4762-43c6-b1b7-906e670217ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b3642617df8fcafa59d2eb42352cfe1f5a455e0a702843c2fffc8491cde5c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b888a1477657d057e486b872a2a3da23225070bfee8b34139b33f4762c4d32fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e69e47ea9ea59fe65c19fad4217224390526bb50c23984de97db2cefa6cfe0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.255034 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.275118 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.284247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.284290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.284302 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.284321 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.284334 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:45Z","lastTransitionTime":"2025-11-24T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.290954 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.307662 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.320994 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.340906 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.358751 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:44Z\\\",\\\"message\\\":\\\"2025-11-24T06:47:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8\\\\n2025-11-24T06:47:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8 to /host/opt/cni/bin/\\\\n2025-11-24T06:47:59Z [verbose] multus-daemon started\\\\n2025-11-24T06:47:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:48:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.376191 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.387245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.387298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.387311 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.387334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.387348 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:45Z","lastTransitionTime":"2025-11-24T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.397126 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.420612 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"t handler 2 for removal\\\\nI1124 06:48:22.681331 6439 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:22.681337 6439 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681376 6439 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 06:48:22.681251 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681402 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 06:48:22.681399 6439 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681604 6439 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:48:22.681615 6439 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681776 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681629 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:22.682681 6439 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.435361 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.449309 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.465099 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.480507 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.490472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.490543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.490567 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.490596 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.490614 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:45Z","lastTransitionTime":"2025-11-24T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.496873 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.514498 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.531966 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.593819 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.593892 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.593903 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.593923 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.593933 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:45Z","lastTransitionTime":"2025-11-24T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.645006 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.672105 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.697071 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.697130 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.697147 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.697178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.697193 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:45Z","lastTransitionTime":"2025-11-24T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.697362 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:44Z\\\",\\\"message\\\":\\\"2025-11-24T06:47:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8\\\\n2025-11-24T06:47:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8 to /host/opt/cni/bin/\\\\n2025-11-24T06:47:59Z [verbose] multus-daemon started\\\\n2025-11-24T06:47:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:48:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.722386 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.743877 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.763225 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.792521 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.800345 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.800378 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.800387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.800407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.800439 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:45Z","lastTransitionTime":"2025-11-24T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.813456 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.840069 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"t handler 2 for removal\\\\nI1124 06:48:22.681331 6439 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:22.681337 6439 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681376 6439 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 06:48:22.681251 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681402 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 06:48:22.681399 6439 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681604 6439 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:48:22.681615 6439 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681776 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681629 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:22.682681 6439 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.857710 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.873100 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.889870 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.901723 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.904234 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.904306 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.904325 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.904349 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.904368 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:45Z","lastTransitionTime":"2025-11-24T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.919226 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.933099 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.946715 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.960467 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ab620d5-4762-43c6-b1b7-906e670217ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b3642617df8fcafa59d2eb42352cfe1f5a455e0a702843c2fffc8491cde5c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b888a1477657d057e486b872a2a3da23225070bfee8b34139b33f4762c4d32fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e69e47ea9ea59fe65c19fad4217224390526bb50c23984de97db2cefa6cfe0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:45 crc kubenswrapper[4799]: I1124 06:48:45.992414 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:45Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.007386 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.007425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.007436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.007451 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.007462 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:46Z","lastTransitionTime":"2025-11-24T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.110131 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.110228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.110253 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.110401 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.110442 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:46Z","lastTransitionTime":"2025-11-24T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.213371 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.213453 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.213470 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.213496 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.213512 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:46Z","lastTransitionTime":"2025-11-24T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.221655 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvd7f_7980021c-dde3-4c14-a7b6-fbcc947a183d/kube-multus/0.log" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.221754 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvd7f" event={"ID":"7980021c-dde3-4c14-a7b6-fbcc947a183d","Type":"ContainerStarted","Data":"ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365"} Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.243656 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.262356 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:44Z\\\",\\\"message\\\":\\\"2025-11-24T06:47:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8\\\\n2025-11-24T06:47:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8 to /host/opt/cni/bin/\\\\n2025-11-24T06:47:59Z [verbose] multus-daemon started\\\\n2025-11-24T06:47:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:48:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.280885 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.296004 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.312506 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.317746 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.317813 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.317836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.317896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.317915 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:46Z","lastTransitionTime":"2025-11-24T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.329579 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.344654 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.360951 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.384280 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"t handler 2 for removal\\\\nI1124 06:48:22.681331 6439 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:22.681337 6439 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681376 6439 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 06:48:22.681251 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681402 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 06:48:22.681399 6439 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681604 6439 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:48:22.681615 6439 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681776 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681629 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:22.682681 6439 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.400329 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.415522 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.420230 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.420305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.420325 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.420355 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.420374 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:46Z","lastTransitionTime":"2025-11-24T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.428542 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.441091 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.454142 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.467653 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.478024 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.488666 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ab620d5-4762-43c6-b1b7-906e670217ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b3642617df8fcafa59d2eb42352cfe1f5a455e0a702843c2fffc8491cde5c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b888a1477657d057e486b872a2a3da23225070bfee8b34139b33f4762c4d32fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e69e47ea9ea59fe65c19fad4217224390526bb50c23984de97db2cefa6cfe0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.506696 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:46Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.523008 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.523053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.523069 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.523087 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.523100 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:46Z","lastTransitionTime":"2025-11-24T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.625869 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.626276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.626353 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.626455 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.627094 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:46Z","lastTransitionTime":"2025-11-24T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.627386 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.627401 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.627504 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:46 crc kubenswrapper[4799]: E1124 06:48:46.627574 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.627675 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:46 crc kubenswrapper[4799]: E1124 06:48:46.627809 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:46 crc kubenswrapper[4799]: E1124 06:48:46.628067 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:46 crc kubenswrapper[4799]: E1124 06:48:46.628174 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.730721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.730779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.730791 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.730811 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.730825 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:46Z","lastTransitionTime":"2025-11-24T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.833828 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.833914 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.833933 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.833960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.833979 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:46Z","lastTransitionTime":"2025-11-24T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.937218 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.937290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.937309 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.937335 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:46 crc kubenswrapper[4799]: I1124 06:48:46.937354 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:46Z","lastTransitionTime":"2025-11-24T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.040106 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.040165 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.040177 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.040199 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.040213 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:47Z","lastTransitionTime":"2025-11-24T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.143269 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.143344 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.143366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.143391 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.143409 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:47Z","lastTransitionTime":"2025-11-24T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.246817 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.246970 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.246989 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.247018 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.247038 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:47Z","lastTransitionTime":"2025-11-24T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.349971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.350354 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.350442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.350526 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.350609 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:47Z","lastTransitionTime":"2025-11-24T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.453925 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.453982 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.453999 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.454025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.454045 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:47Z","lastTransitionTime":"2025-11-24T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.557478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.557564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.557587 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.557614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.557632 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:47Z","lastTransitionTime":"2025-11-24T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.659980 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.660021 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.660031 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.660050 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.660063 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:47Z","lastTransitionTime":"2025-11-24T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.762798 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.762901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.762918 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.762942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.762965 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:47Z","lastTransitionTime":"2025-11-24T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.868135 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.868233 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.868251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.868278 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.868299 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:47Z","lastTransitionTime":"2025-11-24T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.972474 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.972582 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.972609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.972646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:47 crc kubenswrapper[4799]: I1124 06:48:47.972669 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:47Z","lastTransitionTime":"2025-11-24T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.076218 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.076306 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.076328 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.076539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.076561 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:48Z","lastTransitionTime":"2025-11-24T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.180932 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.180988 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.181000 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.181019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.181032 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:48Z","lastTransitionTime":"2025-11-24T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.284658 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.284722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.284736 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.284762 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.284778 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:48Z","lastTransitionTime":"2025-11-24T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.387926 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.387974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.387987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.388011 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.388030 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:48Z","lastTransitionTime":"2025-11-24T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.490390 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.490422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.490432 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.490449 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.490460 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:48Z","lastTransitionTime":"2025-11-24T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.593263 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.593329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.593341 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.593361 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.593374 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:48Z","lastTransitionTime":"2025-11-24T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.628070 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.628095 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.628070 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.628173 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:48 crc kubenswrapper[4799]: E1124 06:48:48.628276 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:48 crc kubenswrapper[4799]: E1124 06:48:48.628213 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:48 crc kubenswrapper[4799]: E1124 06:48:48.628418 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:48 crc kubenswrapper[4799]: E1124 06:48:48.628512 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.696197 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.696245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.696256 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.696275 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.696288 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:48Z","lastTransitionTime":"2025-11-24T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.800392 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.800475 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.800495 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.800521 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.800539 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:48Z","lastTransitionTime":"2025-11-24T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.903174 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.903724 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.903754 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.903791 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:48 crc kubenswrapper[4799]: I1124 06:48:48.903834 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:48Z","lastTransitionTime":"2025-11-24T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.013060 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.014660 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.014832 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.015026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.015176 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:49Z","lastTransitionTime":"2025-11-24T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.119703 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.120246 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.120478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.120676 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.120922 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:49Z","lastTransitionTime":"2025-11-24T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.224935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.225311 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.225503 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.225644 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.225793 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:49Z","lastTransitionTime":"2025-11-24T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.329619 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.329691 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.329708 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.329733 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.329751 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:49Z","lastTransitionTime":"2025-11-24T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.433208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.433290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.433309 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.433336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.433355 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:49Z","lastTransitionTime":"2025-11-24T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.536399 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.536469 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.536489 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.536516 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.536537 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:49Z","lastTransitionTime":"2025-11-24T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.639136 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.639200 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.639222 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.639249 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.639266 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:49Z","lastTransitionTime":"2025-11-24T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.743040 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.743105 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.743127 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.743154 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.743174 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:49Z","lastTransitionTime":"2025-11-24T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.846344 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.846433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.846452 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.846481 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.846506 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:49Z","lastTransitionTime":"2025-11-24T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.950006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.950322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.950342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.950365 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:49 crc kubenswrapper[4799]: I1124 06:48:49.950380 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:49Z","lastTransitionTime":"2025-11-24T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.052702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.052744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.052754 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.052769 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.052780 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:50Z","lastTransitionTime":"2025-11-24T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.155767 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.155836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.155912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.155939 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.155958 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:50Z","lastTransitionTime":"2025-11-24T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.259402 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.259504 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.259531 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.259570 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.259595 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:50Z","lastTransitionTime":"2025-11-24T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.363415 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.363501 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.363529 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.363567 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.363594 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:50Z","lastTransitionTime":"2025-11-24T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.467129 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.467175 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.467188 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.467208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.467225 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:50Z","lastTransitionTime":"2025-11-24T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.570169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.570237 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.570257 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.570286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.570308 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:50Z","lastTransitionTime":"2025-11-24T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.627675 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.627815 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:50 crc kubenswrapper[4799]: E1124 06:48:50.627975 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.627684 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.627728 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:50 crc kubenswrapper[4799]: E1124 06:48:50.628139 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:50 crc kubenswrapper[4799]: E1124 06:48:50.628314 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:50 crc kubenswrapper[4799]: E1124 06:48:50.628552 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.673781 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.673839 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.673896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.673922 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.673941 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:50Z","lastTransitionTime":"2025-11-24T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.777447 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.777518 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.777538 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.777569 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.777590 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:50Z","lastTransitionTime":"2025-11-24T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.881786 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.881875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.881890 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.881914 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.881929 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:50Z","lastTransitionTime":"2025-11-24T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.985787 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.985922 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.985943 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.985971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:50 crc kubenswrapper[4799]: I1124 06:48:50.985988 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:50Z","lastTransitionTime":"2025-11-24T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.089664 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.089816 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.089881 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.089922 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.089945 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:51Z","lastTransitionTime":"2025-11-24T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.195297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.195357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.195375 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.195400 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.195418 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:51Z","lastTransitionTime":"2025-11-24T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.298998 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.299061 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.299071 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.299091 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.299103 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:51Z","lastTransitionTime":"2025-11-24T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.402734 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.402810 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.402833 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.402887 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.402909 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:51Z","lastTransitionTime":"2025-11-24T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.505879 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.505956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.505974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.506006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.506029 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:51Z","lastTransitionTime":"2025-11-24T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.610420 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.610509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.610535 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.610573 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.610602 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:51Z","lastTransitionTime":"2025-11-24T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.714210 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.714279 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.714298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.714330 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.714352 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:51Z","lastTransitionTime":"2025-11-24T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.817717 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.817789 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.817812 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.817879 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.817908 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:51Z","lastTransitionTime":"2025-11-24T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.921755 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.921840 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.921894 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.921927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:51 crc kubenswrapper[4799]: I1124 06:48:51.921947 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:51Z","lastTransitionTime":"2025-11-24T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.025621 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.025687 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.025705 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.025732 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.025752 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:52Z","lastTransitionTime":"2025-11-24T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.129453 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.129542 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.129562 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.129587 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.129606 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:52Z","lastTransitionTime":"2025-11-24T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.234193 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.234293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.234320 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.234361 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.234387 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:52Z","lastTransitionTime":"2025-11-24T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.338123 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.338579 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.338737 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.338937 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.339130 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:52Z","lastTransitionTime":"2025-11-24T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.442584 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.443123 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.443323 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.443512 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.443696 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:52Z","lastTransitionTime":"2025-11-24T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.547669 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.547746 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.547766 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.547797 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.547819 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:52Z","lastTransitionTime":"2025-11-24T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.627996 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.628017 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.628112 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:52 crc kubenswrapper[4799]: E1124 06:48:52.628238 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.628283 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:52 crc kubenswrapper[4799]: E1124 06:48:52.628549 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:52 crc kubenswrapper[4799]: E1124 06:48:52.628625 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:52 crc kubenswrapper[4799]: E1124 06:48:52.628749 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.652128 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.652200 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.652222 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.652306 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.652341 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:52Z","lastTransitionTime":"2025-11-24T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.756155 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.756236 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.756251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.756274 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.756289 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:52Z","lastTransitionTime":"2025-11-24T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.860336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.860427 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.860454 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.860492 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.860513 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:52Z","lastTransitionTime":"2025-11-24T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.963930 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.964011 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.964037 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.964076 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:52 crc kubenswrapper[4799]: I1124 06:48:52.964105 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:52Z","lastTransitionTime":"2025-11-24T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.067184 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.067349 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.067379 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.067414 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.067474 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.170563 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.170646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.170661 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.170709 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.170725 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.274345 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.274429 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.274444 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.274488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.274504 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.294321 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.294387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.294401 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.294422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.294438 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: E1124 06:48:53.312268 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:53Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.317896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.317965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.317988 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.318019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.318042 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: E1124 06:48:53.351244 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:53Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.357430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.357508 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.357535 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.357575 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.357653 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: E1124 06:48:53.380932 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:53Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.387156 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.387209 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.387228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.387256 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.387279 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: E1124 06:48:53.406085 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:53Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.412586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.412668 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.412703 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.412815 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.412885 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: E1124 06:48:53.428682 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:53Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:53 crc kubenswrapper[4799]: E1124 06:48:53.429055 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.431328 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.431396 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.431410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.431436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.431452 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.536264 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.536365 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.536392 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.536430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.536458 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.639905 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.639970 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.639989 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.640014 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.640037 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.742611 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.742680 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.742735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.742768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.742791 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.846395 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.846463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.846491 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.846528 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.846556 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.950156 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.950251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.950284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.950324 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:53 crc kubenswrapper[4799]: I1124 06:48:53.950366 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:53Z","lastTransitionTime":"2025-11-24T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.054159 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.054253 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.054276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.054309 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.054329 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:54Z","lastTransitionTime":"2025-11-24T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.158068 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.158127 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.158139 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.158157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.158172 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:54Z","lastTransitionTime":"2025-11-24T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.261233 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.261298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.261319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.261345 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.261364 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:54Z","lastTransitionTime":"2025-11-24T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.365527 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.365601 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.365626 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.365654 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.365673 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:54Z","lastTransitionTime":"2025-11-24T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.469381 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.469470 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.469489 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.469517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.469538 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:54Z","lastTransitionTime":"2025-11-24T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.572724 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.572812 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.572834 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.572957 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.572980 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:54Z","lastTransitionTime":"2025-11-24T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.628302 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.628397 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.628683 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.628936 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:54 crc kubenswrapper[4799]: E1124 06:48:54.628922 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:54 crc kubenswrapper[4799]: E1124 06:48:54.629158 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.629209 4799 scope.go:117] "RemoveContainer" containerID="dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0" Nov 24 06:48:54 crc kubenswrapper[4799]: E1124 06:48:54.629323 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:54 crc kubenswrapper[4799]: E1124 06:48:54.629458 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.648251 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.676071 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.676138 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.676157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.676185 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.676206 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:54Z","lastTransitionTime":"2025-11-24T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.779145 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.779190 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.779203 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.779220 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.779231 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:54Z","lastTransitionTime":"2025-11-24T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.882621 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.882701 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.882722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.882753 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.882775 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:54Z","lastTransitionTime":"2025-11-24T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.987397 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.987447 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.987458 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.987473 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:54 crc kubenswrapper[4799]: I1124 06:48:54.987486 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:54Z","lastTransitionTime":"2025-11-24T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.090819 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.090941 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.090961 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.090991 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.091013 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:55Z","lastTransitionTime":"2025-11-24T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.194987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.195047 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.195060 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.195085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.195101 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:55Z","lastTransitionTime":"2025-11-24T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.263729 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/2.log" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.267207 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681"} Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.267962 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.280194 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.293208 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.297492 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.297558 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.297577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.297609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.297637 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:55Z","lastTransitionTime":"2025-11-24T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.310551 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.331056 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.348005 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.370455 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6afacfe-8488-4140-a270-7f8fba6ada66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d4ddf8a6fc1ed958f9aa914d2ee1abd73dfe6c36513f0595068a4eea89d6e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.387180 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.400660 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.400697 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.400707 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.400726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.400738 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:55Z","lastTransitionTime":"2025-11-24T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.402671 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.416836 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ab620d5-4762-43c6-b1b7-906e670217ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b3642617df8fcafa59d2eb42352cfe1f5a455e0a702843c2fffc8491cde5c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b888a1477657d057e486b872a2a3da23225070bfee8b34139b33f4762c4d32fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e69e47ea9ea59fe65c19fad4217224390526bb50c23984de97db2cefa6cfe0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.440717 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.457715 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.472770 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.483498 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.496840 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.503392 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.503451 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.503467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.503489 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.503507 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:55Z","lastTransitionTime":"2025-11-24T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.508713 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:44Z\\\",\\\"message\\\":\\\"2025-11-24T06:47:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8\\\\n2025-11-24T06:47:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8 to /host/opt/cni/bin/\\\\n2025-11-24T06:47:59Z [verbose] multus-daemon started\\\\n2025-11-24T06:47:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:48:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.522260 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.536064 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.549116 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.567913 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"t handler 2 for removal\\\\nI1124 06:48:22.681331 6439 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:22.681337 6439 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681376 6439 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 06:48:22.681251 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681402 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 06:48:22.681399 6439 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681604 6439 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:48:22.681615 6439 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681776 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681629 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:22.682681 6439 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.606218 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.606266 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.606278 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.606298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.606311 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:55Z","lastTransitionTime":"2025-11-24T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.645810 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.662030 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.680498 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:44Z\\\",\\\"message\\\":\\\"2025-11-24T06:47:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8\\\\n2025-11-24T06:47:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8 to /host/opt/cni/bin/\\\\n2025-11-24T06:47:59Z [verbose] multus-daemon started\\\\n2025-11-24T06:47:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:48:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.698267 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.709875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.710052 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.710078 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.710129 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.710149 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:55Z","lastTransitionTime":"2025-11-24T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.713730 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.735214 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.750653 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.769190 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.804983 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"t handler 2 for removal\\\\nI1124 06:48:22.681331 6439 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:22.681337 6439 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681376 6439 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 06:48:22.681251 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681402 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 06:48:22.681399 6439 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681604 6439 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:48:22.681615 6439 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681776 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681629 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:22.682681 6439 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.813378 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.813464 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.813481 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.813507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.813527 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:55Z","lastTransitionTime":"2025-11-24T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.828051 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.850242 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.868958 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.884146 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.901705 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.914046 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6afacfe-8488-4140-a270-7f8fba6ada66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d4ddf8a6fc1ed958f9aa914d2ee1abd73dfe6c36513f0595068a4eea89d6e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.916267 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.916293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.916305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.916324 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.916337 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:55Z","lastTransitionTime":"2025-11-24T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.933940 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.951396 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.970539 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ab620d5-4762-43c6-b1b7-906e670217ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b3642617df8fcafa59d2eb42352cfe1f5a455e0a702843c2fffc8491cde5c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b888a1477657d057e486b872a2a3da23225070bfee8b34139b33f4762c4d32fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e69e47ea9ea59fe65c19fad4217224390526bb50c23984de97db2cefa6cfe0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:55 crc kubenswrapper[4799]: I1124 06:48:55.995911 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:55Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.018956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.019007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.019025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.019049 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.019066 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:56Z","lastTransitionTime":"2025-11-24T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.122243 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.122316 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.122336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.122366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.122387 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:56Z","lastTransitionTime":"2025-11-24T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.225443 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.225507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.225522 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.225546 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.225562 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:56Z","lastTransitionTime":"2025-11-24T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.273796 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/3.log" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.274715 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/2.log" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.279745 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" exitCode=1 Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.279812 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681"} Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.279927 4799 scope.go:117] "RemoveContainer" containerID="dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.281437 4799 scope.go:117] "RemoveContainer" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:48:56 crc kubenswrapper[4799]: E1124 06:48:56.282031 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.307876 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.328822 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.328903 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.328921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.328946 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.328966 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:56Z","lastTransitionTime":"2025-11-24T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.329659 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.350420 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.369555 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.387473 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.402208 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6afacfe-8488-4140-a270-7f8fba6ada66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d4ddf8a6fc1ed958f9aa914d2ee1abd73dfe6c36513f0595068a4eea89d6e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.423117 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.433514 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.433582 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.433604 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.433635 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.433655 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:56Z","lastTransitionTime":"2025-11-24T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.447617 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.465153 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ab620d5-4762-43c6-b1b7-906e670217ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b3642617df8fcafa59d2eb42352cfe1f5a455e0a702843c2fffc8491cde5c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b888a1477657d057e486b872a2a3da23225070bfee8b34139b33f4762c4d32fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e69e47ea9ea59fe65c19fad4217224390526bb50c23984de97db2cefa6cfe0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.504737 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.531308 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.536427 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.536463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.536472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.536491 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.536502 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:56Z","lastTransitionTime":"2025-11-24T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.549170 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:44Z\\\",\\\"message\\\":\\\"2025-11-24T06:47:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8\\\\n2025-11-24T06:47:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8 to /host/opt/cni/bin/\\\\n2025-11-24T06:47:59Z [verbose] multus-daemon started\\\\n2025-11-24T06:47:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:48:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.577596 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.628117 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.628185 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.628181 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.628144 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:56 crc kubenswrapper[4799]: E1124 06:48:56.628433 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:56 crc kubenswrapper[4799]: E1124 06:48:56.628537 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:56 crc kubenswrapper[4799]: E1124 06:48:56.628619 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.628862 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.640083 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.640319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.640406 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.640470 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.640525 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:56Z","lastTransitionTime":"2025-11-24T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:56 crc kubenswrapper[4799]: E1124 06:48:56.628323 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.658915 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.669705 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.680627 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.691587 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.710265 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dddc8f757424260607369899a62faa010c70e65a48d64a8b88380fe8b66e7be0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:22Z\\\",\\\"message\\\":\\\"t handler 2 for removal\\\\nI1124 06:48:22.681331 6439 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 06:48:22.681337 6439 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681376 6439 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 06:48:22.681251 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681402 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 06:48:22.681399 6439 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681604 6439 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 06:48:22.681615 6439 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 06:48:22.681776 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 06:48:22.681629 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 06:48:22.682681 6439 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:55Z\\\",\\\"message\\\":\\\"ol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 06:48:55.618722 6806 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 06:48:55.617281 6806 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:56Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.742713 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.742744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.742755 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.742773 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.742783 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:56Z","lastTransitionTime":"2025-11-24T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.845745 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.845806 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.845826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.845874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.845895 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:56Z","lastTransitionTime":"2025-11-24T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.948832 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.948972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.949003 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.949039 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:56 crc kubenswrapper[4799]: I1124 06:48:56.949063 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:56Z","lastTransitionTime":"2025-11-24T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.052962 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.053038 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.053058 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.053088 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.053109 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:57Z","lastTransitionTime":"2025-11-24T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.155979 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.156066 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.156086 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.156123 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.156144 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:57Z","lastTransitionTime":"2025-11-24T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.259564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.259667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.259702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.259750 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.259775 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:57Z","lastTransitionTime":"2025-11-24T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.287571 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/3.log" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.293016 4799 scope.go:117] "RemoveContainer" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:48:57 crc kubenswrapper[4799]: E1124 06:48:57.293235 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.314330 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6afacfe-8488-4140-a270-7f8fba6ada66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d4ddf8a6fc1ed958f9aa914d2ee1abd73dfe6c36513f0595068a4eea89d6e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.334321 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.357233 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.362874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.362957 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.362978 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.363017 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.363063 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:57Z","lastTransitionTime":"2025-11-24T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.372667 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ab620d5-4762-43c6-b1b7-906e670217ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b3642617df8fcafa59d2eb42352cfe1f5a455e0a702843c2fffc8491cde5c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b888a1477657d057e486b872a2a3da23225070bfee8b34139b33f4762c4d32fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e69e47ea9ea59fe65c19fad4217224390526bb50c23984de97db2cefa6cfe0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.391424 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.404324 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:44Z\\\",\\\"message\\\":\\\"2025-11-24T06:47:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8\\\\n2025-11-24T06:47:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8 to /host/opt/cni/bin/\\\\n2025-11-24T06:47:59Z [verbose] multus-daemon started\\\\n2025-11-24T06:47:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:48:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.419576 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.431909 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.448785 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.465668 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.466056 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.466102 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.466111 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.466128 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.466139 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:57Z","lastTransitionTime":"2025-11-24T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.480539 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.499226 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.518983 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.545336 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:55Z\\\",\\\"message\\\":\\\"ol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 06:48:55.618722 6806 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 06:48:55.617281 6806 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.563610 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.569675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.569736 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.569750 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.569777 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.569794 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:57Z","lastTransitionTime":"2025-11-24T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.581402 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.596941 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.612081 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.625934 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:48:57Z is after 2025-08-24T17:21:41Z" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.672880 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.673223 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.673299 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.673370 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.673472 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:57Z","lastTransitionTime":"2025-11-24T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.777608 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.777708 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.777722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.777791 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.777809 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:57Z","lastTransitionTime":"2025-11-24T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.880765 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.880816 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.880830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.880882 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.880898 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:57Z","lastTransitionTime":"2025-11-24T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.984900 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.984984 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.985007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.985043 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:57 crc kubenswrapper[4799]: I1124 06:48:57.985113 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:57Z","lastTransitionTime":"2025-11-24T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.087548 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.087626 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.087652 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.087685 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.087711 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:58Z","lastTransitionTime":"2025-11-24T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.191257 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.191328 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.191344 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.191368 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.191439 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:58Z","lastTransitionTime":"2025-11-24T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.294380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.294436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.294453 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.294478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.294497 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:58Z","lastTransitionTime":"2025-11-24T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.397943 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.398003 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.398017 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.398043 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.398059 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:58Z","lastTransitionTime":"2025-11-24T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.503604 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.503711 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.503731 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.503785 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.503802 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:58Z","lastTransitionTime":"2025-11-24T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.606491 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.606575 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.606602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.606633 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.606659 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:58Z","lastTransitionTime":"2025-11-24T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.628088 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.628123 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:58 crc kubenswrapper[4799]: E1124 06:48:58.628317 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.628367 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.628339 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:48:58 crc kubenswrapper[4799]: E1124 06:48:58.628465 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:48:58 crc kubenswrapper[4799]: E1124 06:48:58.628533 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:48:58 crc kubenswrapper[4799]: E1124 06:48:58.628592 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.710800 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.710918 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.710944 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.710981 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.711007 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:58Z","lastTransitionTime":"2025-11-24T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.813956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.814024 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.814045 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.814074 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.814096 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:58Z","lastTransitionTime":"2025-11-24T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.916504 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.916572 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.916589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.916619 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:58 crc kubenswrapper[4799]: I1124 06:48:58.916648 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:58Z","lastTransitionTime":"2025-11-24T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.019280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.020029 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.020134 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.020178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.020208 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:59Z","lastTransitionTime":"2025-11-24T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.124454 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.124790 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.124954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.125109 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.125267 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:59Z","lastTransitionTime":"2025-11-24T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.228744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.229087 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.229261 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.229397 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.229524 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:59Z","lastTransitionTime":"2025-11-24T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.332270 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.332325 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.332339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.332359 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.332374 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:59Z","lastTransitionTime":"2025-11-24T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.435604 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.435983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.436094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.436237 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.436378 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:59Z","lastTransitionTime":"2025-11-24T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.473965 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.474160 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.474117386 +0000 UTC m=+149.130099900 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.540213 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.540284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.540298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.540319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.540336 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:59Z","lastTransitionTime":"2025-11-24T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.575523 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.575578 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.575606 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.575648 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.575810 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.575836 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.575871 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.575810 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.575941 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.575919274 +0000 UTC m=+149.231901748 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.576014 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.576054 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.575890 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.575828 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.576178 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.576148291 +0000 UTC m=+149.232130775 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.576208 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.576197162 +0000 UTC m=+149.232179656 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 06:48:59 crc kubenswrapper[4799]: E1124 06:48:59.576228 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.576217933 +0000 UTC m=+149.232200417 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.644667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.644774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.644820 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.644922 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.644949 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:59Z","lastTransitionTime":"2025-11-24T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.748727 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.748772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.748784 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.748801 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.748812 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:59Z","lastTransitionTime":"2025-11-24T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.853043 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.853135 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.853163 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.853200 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.853225 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:59Z","lastTransitionTime":"2025-11-24T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.956550 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.956633 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.956646 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.956689 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:48:59 crc kubenswrapper[4799]: I1124 06:48:59.956704 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:48:59Z","lastTransitionTime":"2025-11-24T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.059752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.059815 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.059828 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.059873 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.059889 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:00Z","lastTransitionTime":"2025-11-24T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.163628 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.163696 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.163715 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.163741 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.163762 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:00Z","lastTransitionTime":"2025-11-24T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.267785 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.267900 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.267937 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.267969 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.267993 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:00Z","lastTransitionTime":"2025-11-24T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.371259 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.371336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.371362 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.371397 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.371421 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:00Z","lastTransitionTime":"2025-11-24T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.475357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.475429 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.475450 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.475476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.475494 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:00Z","lastTransitionTime":"2025-11-24T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.579350 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.579410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.579428 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.579454 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.579474 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:00Z","lastTransitionTime":"2025-11-24T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.628290 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.628333 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.628472 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.628596 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:00 crc kubenswrapper[4799]: E1124 06:49:00.628902 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:00 crc kubenswrapper[4799]: E1124 06:49:00.629156 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:00 crc kubenswrapper[4799]: E1124 06:49:00.629292 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:00 crc kubenswrapper[4799]: E1124 06:49:00.629491 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.683242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.683308 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.683319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.683336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.683346 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:00Z","lastTransitionTime":"2025-11-24T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.786310 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.786377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.786405 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.786435 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.786456 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:00Z","lastTransitionTime":"2025-11-24T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.890222 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.890270 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.890278 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.890299 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.890312 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:00Z","lastTransitionTime":"2025-11-24T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.992906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.992965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.992976 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.992996 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:00 crc kubenswrapper[4799]: I1124 06:49:00.993012 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:00Z","lastTransitionTime":"2025-11-24T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.096437 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.096490 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.096507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.096531 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.096545 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:01Z","lastTransitionTime":"2025-11-24T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.199761 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.199801 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.199809 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.199825 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.199837 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:01Z","lastTransitionTime":"2025-11-24T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.302820 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.302914 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.302931 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.302959 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.302977 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:01Z","lastTransitionTime":"2025-11-24T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.406328 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.406412 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.406436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.406467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.406491 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:01Z","lastTransitionTime":"2025-11-24T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.510191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.510247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.510267 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.510290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.510307 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:01Z","lastTransitionTime":"2025-11-24T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.613486 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.613560 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.613580 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.613610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.613630 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:01Z","lastTransitionTime":"2025-11-24T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.717258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.717321 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.717335 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.717357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.717371 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:01Z","lastTransitionTime":"2025-11-24T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.821359 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.821434 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.821452 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.821481 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.821506 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:01Z","lastTransitionTime":"2025-11-24T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.924169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.924217 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.924234 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.924257 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:01 crc kubenswrapper[4799]: I1124 06:49:01.924272 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:01Z","lastTransitionTime":"2025-11-24T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.027764 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.027826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.027872 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.027938 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.027961 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:02Z","lastTransitionTime":"2025-11-24T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.131573 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.131636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.131658 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.131690 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.131716 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:02Z","lastTransitionTime":"2025-11-24T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.235189 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.235252 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.235271 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.235301 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.235329 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:02Z","lastTransitionTime":"2025-11-24T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.343168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.343273 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.343295 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.343335 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.343355 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:02Z","lastTransitionTime":"2025-11-24T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.448242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.448393 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.448425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.448465 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.448496 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:02Z","lastTransitionTime":"2025-11-24T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.552524 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.552603 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.552622 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.552653 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.552677 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:02Z","lastTransitionTime":"2025-11-24T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.627836 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.627931 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.627978 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.627977 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:02 crc kubenswrapper[4799]: E1124 06:49:02.628114 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:02 crc kubenswrapper[4799]: E1124 06:49:02.628289 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:02 crc kubenswrapper[4799]: E1124 06:49:02.628465 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:02 crc kubenswrapper[4799]: E1124 06:49:02.628620 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.658131 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.658190 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.658208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.658236 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.658254 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:02Z","lastTransitionTime":"2025-11-24T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.761602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.761661 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.761679 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.761709 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.761729 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:02Z","lastTransitionTime":"2025-11-24T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.865349 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.865423 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.865442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.865468 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.865492 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:02Z","lastTransitionTime":"2025-11-24T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.968607 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.968664 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.968683 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.968708 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:02 crc kubenswrapper[4799]: I1124 06:49:02.968728 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:02Z","lastTransitionTime":"2025-11-24T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.072623 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.072701 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.072721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.072749 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.072769 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.176283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.176374 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.176397 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.176455 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.176474 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.280272 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.280357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.280380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.280409 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.280427 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.384585 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.384655 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.384675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.384703 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.384724 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.488206 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.488269 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.488289 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.488319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.488339 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.592007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.592077 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.592101 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.592133 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.592156 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.695886 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.695978 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.695999 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.696027 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.696046 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.766145 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.766225 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.766243 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.766275 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.766296 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: E1124 06:49:03.789488 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.795599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.795682 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.795709 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.795749 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.795770 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: E1124 06:49:03.818718 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.824933 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.825020 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.825046 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.825080 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.825105 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: E1124 06:49:03.845200 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.850672 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.850798 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.850827 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.850923 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.850943 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: E1124 06:49:03.873267 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.878564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.878599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.878607 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.878622 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.878631 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:03 crc kubenswrapper[4799]: E1124 06:49:03.894247 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T06:49:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5d29851c-ec7b-44cd-80ee-83d663957897\\\",\\\"systemUUID\\\":\\\"8403adc0-fd03-431c-8d73-b29da3901e7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:03Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:03 crc kubenswrapper[4799]: E1124 06:49:03.894386 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.896493 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.896526 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.896538 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.896575 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:03 crc kubenswrapper[4799]: I1124 06:49:03.896591 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:03Z","lastTransitionTime":"2025-11-24T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:03.999967 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.000066 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.000091 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.000130 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.000155 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:04Z","lastTransitionTime":"2025-11-24T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.104615 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.104700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.104720 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.104747 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.104767 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:04Z","lastTransitionTime":"2025-11-24T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.209039 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.209130 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.209151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.209182 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.209202 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:04Z","lastTransitionTime":"2025-11-24T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.313449 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.313521 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.313539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.313566 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.313587 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:04Z","lastTransitionTime":"2025-11-24T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.416664 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.416730 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.416748 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.416775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.416910 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:04Z","lastTransitionTime":"2025-11-24T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.520500 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.520599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.520618 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.520648 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.520670 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:04Z","lastTransitionTime":"2025-11-24T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.624293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.624357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.624372 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.624393 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.624406 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:04Z","lastTransitionTime":"2025-11-24T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.627668 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.627786 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:04 crc kubenswrapper[4799]: E1124 06:49:04.627818 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.627842 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.627668 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:04 crc kubenswrapper[4799]: E1124 06:49:04.627973 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:04 crc kubenswrapper[4799]: E1124 06:49:04.628104 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:04 crc kubenswrapper[4799]: E1124 06:49:04.628202 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.727477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.727527 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.727539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.727560 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.727574 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:04Z","lastTransitionTime":"2025-11-24T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.831540 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.832086 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.832102 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.832126 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.832143 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:04Z","lastTransitionTime":"2025-11-24T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.935375 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.935428 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.935444 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.935466 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:04 crc kubenswrapper[4799]: I1124 06:49:04.935480 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:04Z","lastTransitionTime":"2025-11-24T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.038968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.039049 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.039067 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.039093 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.039112 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:05Z","lastTransitionTime":"2025-11-24T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.143051 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.143118 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.143137 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.143167 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.143186 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:05Z","lastTransitionTime":"2025-11-24T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.247680 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.247761 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.247790 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.247830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.247907 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:05Z","lastTransitionTime":"2025-11-24T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.351630 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.351692 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.351712 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.351737 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.351758 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:05Z","lastTransitionTime":"2025-11-24T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.456152 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.456226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.456245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.456273 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.456293 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:05Z","lastTransitionTime":"2025-11-24T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.560023 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.560088 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.560103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.560124 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.560138 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:05Z","lastTransitionTime":"2025-11-24T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.645282 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6afacfe-8488-4140-a270-7f8fba6ada66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d4ddf8a6fc1ed958f9aa914d2ee1abd73dfe6c36513f0595068a4eea89d6e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92e5b39eeed72d7da9927c226aaa6eae43fb36ad53ae1c2c8e869fe92fc5c508\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.663739 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.663819 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.663842 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.663910 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.663934 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:05Z","lastTransitionTime":"2025-11-24T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.665635 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9295e20f5c0967a9843ab116fc2579551b981fa60ee2ba9acc96c11db7d7f9a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5797cc2e61a49960a21ce1de33ccf17a0f45b76885cbbccd0447da762629a6ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.685127 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c56a6245-1460-4588-b33c-188991e12266\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711dba828b58c4ae96e6683c655a9213d4a7e6958b53d96da7b2116ee44e6bad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e6a7219a04e9a39be12f02b478522d5cd221c16fa7b93ed0774092bb693a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xktbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2vq5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.704832 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ab620d5-4762-43c6-b1b7-906e670217ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b3642617df8fcafa59d2eb42352cfe1f5a455e0a702843c2fffc8491cde5c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b888a1477657d057e486b872a2a3da23225070bfee8b34139b33f4762c4d32fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e69e47ea9ea59fe65c19fad4217224390526bb50c23984de97db2cefa6cfe0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98f4806e1f911c30db8ed74896c13921c23f035a23ff1dd8df0bc727366d410d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.744226 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1541111f-7126-4e3b-923d-b8e8149394c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://759e7038f96e491451e7fe447eb14a7c5b2b5375e03e5e529a486596de6a040b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3736c9f58b67f06c7444895aa83451c507988f2bb4934a46079f2ccbe2520329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e0d6705ca0642e8262b0f4a9f70161fc2a8b0938bcdf055da6883981bb96ac1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e0b688cf331931f090ce91e836311f78567175626632cb4164edaf614cb1b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://866a840c8b95df959d17de9cdef51f92bfe19e4ff9d1b6ba04feb46deacf8b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4568a08b2e7c2ea4a9436081cd58288a311176bc7d7f7d0582373db87fddb9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://930d7b4b89832dd6cf9eba1fb0e614842ec49a22c1ab2c574827c61057b72921\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5557a68819428fef7a4226a07ffdd0078b39f0c4b4329128589ed0418edd643c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.765518 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483069c2f6e2591d5e0f882357f52579f6719a1551d713a10db25e2e4c24c938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.767199 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.767272 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.767295 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.767322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.767343 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:05Z","lastTransitionTime":"2025-11-24T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.787291 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.807439 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bkb8p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2ca5e21-6366-493a-b14b-b71baa6b54d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c31da0b80e5b731079a3948d6a69613b1cebad8599efe581b524f3430e482e95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pq4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bkb8p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.832134 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed80eb00-6567-4878-91c2-979bb8e2f75c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01623e760cd27ff9930346ea54bca0068f28c45e5cfc32d9143b40951e38e111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://496c18d3f79323905c767a8b5b605d75080689f894e32cf600f93ec556e5b42e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ae3ad4087a748f5ab01cc41fa3cf883b5661d7aa57b59aeb978ee51d0c080e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c4e5b29538a49fbcfe6ffe454b8aeb8ef1107b274d8ebef4b87917140e2aabe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e10af32bceac8cfe62bd543af64ac3fae17aefe06969c203c3ccd30cb6b34b44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3757c625122073517f5d3818f3005021385bb0653c7fa6b373a82493606ae46c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4df5b03f9b851ac144d6157c99ccb7a2ddb1ab0b1bbd8104ae39e8e044aa67c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:48:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5ttq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zzvdv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.858734 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvd7f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7980021c-dde3-4c14-a7b6-fbcc947a183d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:44Z\\\",\\\"message\\\":\\\"2025-11-24T06:47:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8\\\\n2025-11-24T06:47:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8b23b60e-d4c8-4f8e-9a3b-333fc292dcd8 to /host/opt/cni/bin/\\\\n2025-11-24T06:47:59Z [verbose] multus-daemon started\\\\n2025-11-24T06:47:59Z [verbose] Readiness Indicator file check\\\\n2025-11-24T06:48:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz6b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvd7f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.875544 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.875621 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.875642 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.875674 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.875697 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:05Z","lastTransitionTime":"2025-11-24T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.890400 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bddb1822-b307-4c00-85dd-4a3259293332\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4be4d414f3d25c79ccd5a9e3fbef831a563ca603000cf63e46437e64f7eeebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d5bc600182cd3e7003c85a61955fd16357a069ca371518b8d7e42bca0dac0b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b72fb7ea2740fcc262b7899fe6bcdea0469cb9621d190e95cf3dfee92d3721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900e4ed2aa16d980a605db02c60638eaf154896554ad825596ad404fd8fdac2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e4df8216246fdbeb0b6cfcef82ca8e67afdbeed070c9b94cc92550505fde715\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1124 06:47:49.243056 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 06:47:49.246189 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-139352087/tls.crt::/tmp/serving-cert-139352087/tls.key\\\\\\\"\\\\nI1124 06:47:55.630139 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 06:47:55.643195 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 06:47:55.643227 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 06:47:55.643260 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 06:47:55.643275 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 06:47:55.660890 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 06:47:55.660948 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660962 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 06:47:55.660975 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 06:47:55.660984 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 06:47:55.660993 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 06:47:55.661001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 06:47:55.661351 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 06:47:55.664282 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f04ec536692056f32fc21001f9c9c989c4ef49f905498218bf1043d27c46050\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9901b62a2ea097509a96baf16534dce3134a9b29e31f4c516e1dfebc202509dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.914405 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cbe4c821-42b0-4111-a21a-d2375cfbf414\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11ad421d0972fc33ab9602f19b92d5287cdf04ea0bbad69be495493726baaae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d9a5af1a34f6597208022153d4ecd0adc7577e9f3233da609a2aa2e3d5c048c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://461611ab012b9459b8fc54f656c8829693b4edf0bf0a35bcc5ab980e1031fadb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef462e9994f3c00c5dc05ca967019da96dfe890a477ea2494bd18a21142a9b0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.936728 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abac17d624f8788d7d1db4e54703a8ac2dc0c3f694c97553b40e4b62dee10b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.972040 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T06:48:55Z\\\",\\\"message\\\":\\\"ol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler/scheduler\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.169\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1124 06:48:55.618722 6806 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-scheduler/scheduler]} name:Service_openshift-kube-scheduler/scheduler_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.169:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {39432221-5995-412b-967b-35e1a9405ec7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 06:48:55.617281 6806 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T06:48:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r8djp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b7nd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.978569 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.978618 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.978634 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.978660 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.978678 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:05Z","lastTransitionTime":"2025-11-24T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:05 crc kubenswrapper[4799]: I1124 06:49:05.987971 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fzb2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2af19-8c5a-454d-a9b1-7f9ae60a9e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292b7f7d8d4e86d6c1965d00307698d9c2fc968b0d49b0cc3a4b94fc011b70d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmxgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fzb2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:05Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.003003 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f587ff07-479b-4c95-95cb-406faffcfb68\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:48:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c87kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:48:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ldbr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.022790 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.043658 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.064071 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"354ebd50-7a67-481d-86db-4b3bf4753161\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T06:47:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58da8119cff5d52d7af403a4eded08139159922579e4b6d125ff53f46a2fb488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T06:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxr2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T06:47:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rzk7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T06:49:06Z is after 2025-08-24T17:21:41Z" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.081753 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.081818 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.081835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.081896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.081951 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:06Z","lastTransitionTime":"2025-11-24T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.185155 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.185385 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.185428 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.185530 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.185644 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:06Z","lastTransitionTime":"2025-11-24T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.289702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.289765 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.289775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.289791 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.289801 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:06Z","lastTransitionTime":"2025-11-24T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.393611 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.393678 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.393697 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.393725 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.393746 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:06Z","lastTransitionTime":"2025-11-24T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.497805 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.497920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.497947 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.497979 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.498004 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:06Z","lastTransitionTime":"2025-11-24T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.601122 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.601176 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.601193 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.601218 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.601235 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:06Z","lastTransitionTime":"2025-11-24T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.627575 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:06 crc kubenswrapper[4799]: E1124 06:49:06.627761 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.628036 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.628083 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:06 crc kubenswrapper[4799]: E1124 06:49:06.628262 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.628280 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:06 crc kubenswrapper[4799]: E1124 06:49:06.628762 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:06 crc kubenswrapper[4799]: E1124 06:49:06.628675 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.704110 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.704177 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.704198 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.704227 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.704246 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:06Z","lastTransitionTime":"2025-11-24T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.808268 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.808347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.808366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.808394 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.808415 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:06Z","lastTransitionTime":"2025-11-24T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.911721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.911902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.911931 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.911999 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:06 crc kubenswrapper[4799]: I1124 06:49:06.912026 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:06Z","lastTransitionTime":"2025-11-24T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.016229 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.016289 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.016303 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.016329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.016346 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:07Z","lastTransitionTime":"2025-11-24T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.119159 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.119235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.119253 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.119286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.119311 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:07Z","lastTransitionTime":"2025-11-24T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.222723 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.222779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.222793 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.222813 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.222826 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:07Z","lastTransitionTime":"2025-11-24T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.325897 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.325998 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.326023 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.326142 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.326198 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:07Z","lastTransitionTime":"2025-11-24T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.429518 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.429572 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.429590 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.429613 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.429634 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:07Z","lastTransitionTime":"2025-11-24T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.533225 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.533291 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.533322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.533357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.533376 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:07Z","lastTransitionTime":"2025-11-24T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.642149 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.642226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.642251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.642281 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.642305 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:07Z","lastTransitionTime":"2025-11-24T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.746385 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.746471 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.746495 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.746527 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.746551 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:07Z","lastTransitionTime":"2025-11-24T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.850814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.851295 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.851502 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.851721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.851970 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:07Z","lastTransitionTime":"2025-11-24T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.956102 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.956214 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.956240 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.956277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:07 crc kubenswrapper[4799]: I1124 06:49:07.956301 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:07Z","lastTransitionTime":"2025-11-24T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.060927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.061422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.061569 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.061718 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.061983 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:08Z","lastTransitionTime":"2025-11-24T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.180934 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.181006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.181026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.181058 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.181080 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:08Z","lastTransitionTime":"2025-11-24T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.284151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.284598 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.284783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.285005 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.285175 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:08Z","lastTransitionTime":"2025-11-24T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.388406 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.389258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.389406 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.389539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.389663 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:08Z","lastTransitionTime":"2025-11-24T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.493518 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.493923 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.494032 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.494151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.494254 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:08Z","lastTransitionTime":"2025-11-24T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.598153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.598596 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.598873 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.599094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.599249 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:08Z","lastTransitionTime":"2025-11-24T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.627662 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.627806 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:08 crc kubenswrapper[4799]: E1124 06:49:08.627920 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.627965 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.627970 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:08 crc kubenswrapper[4799]: E1124 06:49:08.628181 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:08 crc kubenswrapper[4799]: E1124 06:49:08.628297 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:08 crc kubenswrapper[4799]: E1124 06:49:08.628550 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.629725 4799 scope.go:117] "RemoveContainer" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:49:08 crc kubenswrapper[4799]: E1124 06:49:08.630173 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.702429 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.702489 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.702500 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.702524 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.702539 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:08Z","lastTransitionTime":"2025-11-24T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.806953 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.807026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.807044 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.807071 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.807090 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:08Z","lastTransitionTime":"2025-11-24T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.911062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.911152 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.911178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.911212 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:08 crc kubenswrapper[4799]: I1124 06:49:08.911236 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:08Z","lastTransitionTime":"2025-11-24T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.014958 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.015035 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.015070 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.015101 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.015114 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:09Z","lastTransitionTime":"2025-11-24T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.118974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.119052 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.119069 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.119097 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.119116 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:09Z","lastTransitionTime":"2025-11-24T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.223997 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.224120 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.224152 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.224193 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.224235 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:09Z","lastTransitionTime":"2025-11-24T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.327962 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.328043 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.328083 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.328122 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.328146 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:09Z","lastTransitionTime":"2025-11-24T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.432007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.432083 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.432108 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.432139 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.432163 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:09Z","lastTransitionTime":"2025-11-24T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.535190 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.535275 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.535299 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.535326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.535354 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:09Z","lastTransitionTime":"2025-11-24T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.638636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.638696 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.638716 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.638738 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.638756 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:09Z","lastTransitionTime":"2025-11-24T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.742959 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.743045 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.743064 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.743092 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.743115 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:09Z","lastTransitionTime":"2025-11-24T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.846603 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.846701 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.846724 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.846762 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.846789 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:09Z","lastTransitionTime":"2025-11-24T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.950917 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.951054 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.951083 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.951120 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:09 crc kubenswrapper[4799]: I1124 06:49:09.951150 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:09Z","lastTransitionTime":"2025-11-24T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.054991 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.055073 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.055099 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.055135 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.055160 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:10Z","lastTransitionTime":"2025-11-24T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.158454 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.158519 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.158542 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.158570 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.158589 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:10Z","lastTransitionTime":"2025-11-24T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.261753 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.261829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.261878 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.261906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.261927 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:10Z","lastTransitionTime":"2025-11-24T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.366004 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.366081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.366103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.366135 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.366158 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:10Z","lastTransitionTime":"2025-11-24T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.469302 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.469372 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.469388 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.469415 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.469438 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:10Z","lastTransitionTime":"2025-11-24T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.573938 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.574118 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.574154 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.574242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.574316 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:10Z","lastTransitionTime":"2025-11-24T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.627333 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.627384 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.627393 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:10 crc kubenswrapper[4799]: E1124 06:49:10.627560 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.627587 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:10 crc kubenswrapper[4799]: E1124 06:49:10.627763 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:10 crc kubenswrapper[4799]: E1124 06:49:10.627944 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:10 crc kubenswrapper[4799]: E1124 06:49:10.628241 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.678305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.678379 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.678452 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.678497 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.678523 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:10Z","lastTransitionTime":"2025-11-24T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.785537 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.785621 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.785645 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.785696 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.785721 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:10Z","lastTransitionTime":"2025-11-24T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.888578 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.888648 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.888667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.888702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.888728 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:10Z","lastTransitionTime":"2025-11-24T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.992395 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.992472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.992498 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.992533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:10 crc kubenswrapper[4799]: I1124 06:49:10.992557 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:10Z","lastTransitionTime":"2025-11-24T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.096778 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.096871 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.096890 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.096921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.096941 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:11Z","lastTransitionTime":"2025-11-24T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.199918 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.200352 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.200673 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.201017 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.201319 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:11Z","lastTransitionTime":"2025-11-24T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.304519 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.304956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.305198 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.305459 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.305660 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:11Z","lastTransitionTime":"2025-11-24T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.408924 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.409008 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.409030 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.409059 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.409100 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:11Z","lastTransitionTime":"2025-11-24T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.512023 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.512077 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.512094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.512149 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.512179 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:11Z","lastTransitionTime":"2025-11-24T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.617075 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.617166 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.617192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.617228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.617255 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:11Z","lastTransitionTime":"2025-11-24T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.720875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.720941 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.720950 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.720966 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.720975 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:11Z","lastTransitionTime":"2025-11-24T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.824843 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.825351 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.825494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.825641 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.825767 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:11Z","lastTransitionTime":"2025-11-24T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.929095 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.929161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.929179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.929209 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:11 crc kubenswrapper[4799]: I1124 06:49:11.929227 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:11Z","lastTransitionTime":"2025-11-24T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.032952 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.033342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.033572 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.033803 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.034088 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:12Z","lastTransitionTime":"2025-11-24T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.137623 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.137671 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.137682 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.137704 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.137717 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:12Z","lastTransitionTime":"2025-11-24T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.241685 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.241752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.241775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.241805 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.241823 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:12Z","lastTransitionTime":"2025-11-24T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.345544 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.345652 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.345681 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.345752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.345777 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:12Z","lastTransitionTime":"2025-11-24T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.449461 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.449546 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.449567 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.449598 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.449621 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:12Z","lastTransitionTime":"2025-11-24T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.553676 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.553772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.553793 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.553826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.553887 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:12Z","lastTransitionTime":"2025-11-24T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.627368 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.627417 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.627368 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:12 crc kubenswrapper[4799]: E1124 06:49:12.627614 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.627706 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:12 crc kubenswrapper[4799]: E1124 06:49:12.627919 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:12 crc kubenswrapper[4799]: E1124 06:49:12.628065 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:12 crc kubenswrapper[4799]: E1124 06:49:12.628248 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.659521 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.659614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.659653 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.659681 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.659699 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:12Z","lastTransitionTime":"2025-11-24T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.763195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.763318 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.763377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.763411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.763466 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:12Z","lastTransitionTime":"2025-11-24T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.866551 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.866615 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.866632 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.866656 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.866713 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:12Z","lastTransitionTime":"2025-11-24T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.969763 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.969838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.969892 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.969927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:12 crc kubenswrapper[4799]: I1124 06:49:12.969943 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:12Z","lastTransitionTime":"2025-11-24T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.073026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.073085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.073103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.073132 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.073150 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:13Z","lastTransitionTime":"2025-11-24T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.177786 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.177903 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.177923 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.177955 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.177977 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:13Z","lastTransitionTime":"2025-11-24T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.281414 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.281492 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.281512 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.281542 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.281564 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:13Z","lastTransitionTime":"2025-11-24T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.384616 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.384700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.384722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.384758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.384788 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:13Z","lastTransitionTime":"2025-11-24T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.488555 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.488621 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.488640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.488669 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.488693 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:13Z","lastTransitionTime":"2025-11-24T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.592728 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.594005 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.594251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.594493 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.594647 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:13Z","lastTransitionTime":"2025-11-24T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.698274 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.698313 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.698322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.698338 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.698349 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:13Z","lastTransitionTime":"2025-11-24T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.802540 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.802600 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.802619 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.802648 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.802668 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:13Z","lastTransitionTime":"2025-11-24T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.906245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.906307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.906317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.906335 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.906351 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:13Z","lastTransitionTime":"2025-11-24T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.907943 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.908019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.908038 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.908066 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.908087 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T06:49:13Z","lastTransitionTime":"2025-11-24T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.976335 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t"] Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.977206 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.980126 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.980459 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.980690 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 06:49:13 crc kubenswrapper[4799]: I1124 06:49:13.981161 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.021741 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=20.021707272 podStartE2EDuration="20.021707272s" podCreationTimestamp="2025-11-24 06:48:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:14.00198376 +0000 UTC m=+99.657966274" watchObservedRunningTime="2025-11-24 06:49:14.021707272 +0000 UTC m=+99.677689786" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.065807 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.065963 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.066004 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.065986 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2vq5k" podStartSLOduration=77.065957093 podStartE2EDuration="1m17.065957093s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:14.042220598 +0000 UTC m=+99.698203102" watchObservedRunningTime="2025-11-24 06:49:14.065957093 +0000 UTC m=+99.721939607" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.066081 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.066113 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.127128 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.12709883 podStartE2EDuration="45.12709883s" podCreationTimestamp="2025-11-24 06:48:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:14.088713158 +0000 UTC m=+99.744695662" watchObservedRunningTime="2025-11-24 06:49:14.12709883 +0000 UTC m=+99.783081384" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.127498 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=76.127490692 podStartE2EDuration="1m16.127490692s" podCreationTimestamp="2025-11-24 06:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:14.125452349 +0000 UTC m=+99.781434863" watchObservedRunningTime="2025-11-24 06:49:14.127490692 +0000 UTC m=+99.783473196" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.142622 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-zvd7f" podStartSLOduration=78.142595013 podStartE2EDuration="1m18.142595013s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:14.142547581 +0000 UTC m=+99.798530065" watchObservedRunningTime="2025-11-24 06:49:14.142595013 +0000 UTC m=+99.798577487" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.166711 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.166777 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.166797 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.166835 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.166863 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.167113 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.167112 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.168960 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.170656 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=79.170631498 podStartE2EDuration="1m19.170631498s" podCreationTimestamp="2025-11-24 06:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:14.16970478 +0000 UTC m=+99.825687284" watchObservedRunningTime="2025-11-24 06:49:14.170631498 +0000 UTC m=+99.826614012" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.182226 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.191267 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.191237018 podStartE2EDuration="1m18.191237018s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:14.190204806 +0000 UTC m=+99.846187320" watchObservedRunningTime="2025-11-24 06:49:14.191237018 +0000 UTC m=+99.847219522" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.193067 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b8fc4c4c-9ac7-4b08-8759-48db7fafebde-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nvx7t\" (UID: \"b8fc4c4c-9ac7-4b08-8759-48db7fafebde\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.271907 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-bkb8p" podStartSLOduration=78.271834268 podStartE2EDuration="1m18.271834268s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:14.269107075 +0000 UTC m=+99.925089579" watchObservedRunningTime="2025-11-24 06:49:14.271834268 +0000 UTC m=+99.927816832" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.293896 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zzvdv" podStartSLOduration=78.2938393 podStartE2EDuration="1m18.2938393s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:14.293659265 +0000 UTC m=+99.949641779" watchObservedRunningTime="2025-11-24 06:49:14.2938393 +0000 UTC m=+99.949821794" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.296700 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.359077 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" event={"ID":"b8fc4c4c-9ac7-4b08-8759-48db7fafebde","Type":"ContainerStarted","Data":"a1ca2f41976fcf3f6245d54754f6099a18a2da6e8b4d442f4716b7166fddaea4"} Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.392588 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podStartSLOduration=78.392558354 podStartE2EDuration="1m18.392558354s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:14.391013987 +0000 UTC m=+100.046996491" watchObservedRunningTime="2025-11-24 06:49:14.392558354 +0000 UTC m=+100.048540848" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.423052 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-fzb2l" podStartSLOduration=77.423025984 podStartE2EDuration="1m17.423025984s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:14.408258103 +0000 UTC m=+100.064240587" watchObservedRunningTime="2025-11-24 06:49:14.423025984 +0000 UTC m=+100.079008468" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.627223 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.627254 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.627261 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:14 crc kubenswrapper[4799]: E1124 06:49:14.627373 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:14 crc kubenswrapper[4799]: I1124 06:49:14.627382 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:14 crc kubenswrapper[4799]: E1124 06:49:14.627488 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:14 crc kubenswrapper[4799]: E1124 06:49:14.627590 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:14 crc kubenswrapper[4799]: E1124 06:49:14.627809 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:15 crc kubenswrapper[4799]: I1124 06:49:15.179358 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:15 crc kubenswrapper[4799]: E1124 06:49:15.179675 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:49:15 crc kubenswrapper[4799]: E1124 06:49:15.180225 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs podName:f587ff07-479b-4c95-95cb-406faffcfb68 nodeName:}" failed. No retries permitted until 2025-11-24 06:50:19.18018729 +0000 UTC m=+164.836169804 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs") pod "network-metrics-daemon-ldbr4" (UID: "f587ff07-479b-4c95-95cb-406faffcfb68") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 06:49:15 crc kubenswrapper[4799]: I1124 06:49:15.365217 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" event={"ID":"b8fc4c4c-9ac7-4b08-8759-48db7fafebde","Type":"ContainerStarted","Data":"0a320a8e588e840c16a41f2d48d02c13fb6456392c6469539e976d6856035f20"} Nov 24 06:49:16 crc kubenswrapper[4799]: I1124 06:49:16.627371 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:16 crc kubenswrapper[4799]: I1124 06:49:16.627398 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:16 crc kubenswrapper[4799]: I1124 06:49:16.627361 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:16 crc kubenswrapper[4799]: I1124 06:49:16.627490 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:16 crc kubenswrapper[4799]: E1124 06:49:16.627676 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:16 crc kubenswrapper[4799]: E1124 06:49:16.627927 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:16 crc kubenswrapper[4799]: E1124 06:49:16.628117 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:16 crc kubenswrapper[4799]: E1124 06:49:16.628287 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:18 crc kubenswrapper[4799]: I1124 06:49:18.628131 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:18 crc kubenswrapper[4799]: I1124 06:49:18.628199 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:18 crc kubenswrapper[4799]: I1124 06:49:18.628138 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:18 crc kubenswrapper[4799]: I1124 06:49:18.628544 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:18 crc kubenswrapper[4799]: E1124 06:49:18.628762 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:18 crc kubenswrapper[4799]: E1124 06:49:18.629506 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:18 crc kubenswrapper[4799]: E1124 06:49:18.629628 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:18 crc kubenswrapper[4799]: E1124 06:49:18.629388 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:19 crc kubenswrapper[4799]: I1124 06:49:19.642452 4799 scope.go:117] "RemoveContainer" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:49:19 crc kubenswrapper[4799]: E1124 06:49:19.642669 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" Nov 24 06:49:20 crc kubenswrapper[4799]: I1124 06:49:20.628022 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:20 crc kubenswrapper[4799]: I1124 06:49:20.628082 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:20 crc kubenswrapper[4799]: I1124 06:49:20.628100 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:20 crc kubenswrapper[4799]: I1124 06:49:20.628097 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:20 crc kubenswrapper[4799]: E1124 06:49:20.628290 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:20 crc kubenswrapper[4799]: E1124 06:49:20.628419 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:20 crc kubenswrapper[4799]: E1124 06:49:20.628520 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:20 crc kubenswrapper[4799]: E1124 06:49:20.628652 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:22 crc kubenswrapper[4799]: I1124 06:49:22.627512 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:22 crc kubenswrapper[4799]: I1124 06:49:22.627593 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:22 crc kubenswrapper[4799]: I1124 06:49:22.627681 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:22 crc kubenswrapper[4799]: I1124 06:49:22.627750 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:22 crc kubenswrapper[4799]: E1124 06:49:22.627733 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:22 crc kubenswrapper[4799]: E1124 06:49:22.627983 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:22 crc kubenswrapper[4799]: E1124 06:49:22.628183 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:22 crc kubenswrapper[4799]: E1124 06:49:22.628302 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:24 crc kubenswrapper[4799]: I1124 06:49:24.627873 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:24 crc kubenswrapper[4799]: I1124 06:49:24.627928 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:24 crc kubenswrapper[4799]: I1124 06:49:24.627937 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:24 crc kubenswrapper[4799]: I1124 06:49:24.627967 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:24 crc kubenswrapper[4799]: E1124 06:49:24.628170 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:24 crc kubenswrapper[4799]: E1124 06:49:24.628290 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:24 crc kubenswrapper[4799]: E1124 06:49:24.628476 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:24 crc kubenswrapper[4799]: E1124 06:49:24.628589 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:26 crc kubenswrapper[4799]: I1124 06:49:26.628278 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:26 crc kubenswrapper[4799]: I1124 06:49:26.628407 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:26 crc kubenswrapper[4799]: I1124 06:49:26.629165 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:26 crc kubenswrapper[4799]: I1124 06:49:26.629421 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:26 crc kubenswrapper[4799]: E1124 06:49:26.629445 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:26 crc kubenswrapper[4799]: E1124 06:49:26.629562 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:26 crc kubenswrapper[4799]: E1124 06:49:26.629795 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:26 crc kubenswrapper[4799]: E1124 06:49:26.630008 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:28 crc kubenswrapper[4799]: I1124 06:49:28.628513 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:28 crc kubenswrapper[4799]: I1124 06:49:28.628531 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:28 crc kubenswrapper[4799]: E1124 06:49:28.629269 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:28 crc kubenswrapper[4799]: I1124 06:49:28.628583 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:28 crc kubenswrapper[4799]: E1124 06:49:28.629396 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:28 crc kubenswrapper[4799]: I1124 06:49:28.628586 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:28 crc kubenswrapper[4799]: E1124 06:49:28.629669 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:28 crc kubenswrapper[4799]: E1124 06:49:28.629766 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:30 crc kubenswrapper[4799]: I1124 06:49:30.627872 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:30 crc kubenswrapper[4799]: I1124 06:49:30.627920 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:30 crc kubenswrapper[4799]: I1124 06:49:30.627971 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:30 crc kubenswrapper[4799]: E1124 06:49:30.628159 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:30 crc kubenswrapper[4799]: E1124 06:49:30.628347 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:30 crc kubenswrapper[4799]: I1124 06:49:30.628489 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:30 crc kubenswrapper[4799]: E1124 06:49:30.628552 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:30 crc kubenswrapper[4799]: E1124 06:49:30.628663 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:31 crc kubenswrapper[4799]: I1124 06:49:31.429447 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvd7f_7980021c-dde3-4c14-a7b6-fbcc947a183d/kube-multus/1.log" Nov 24 06:49:31 crc kubenswrapper[4799]: I1124 06:49:31.430506 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvd7f_7980021c-dde3-4c14-a7b6-fbcc947a183d/kube-multus/0.log" Nov 24 06:49:31 crc kubenswrapper[4799]: I1124 06:49:31.430587 4799 generic.go:334] "Generic (PLEG): container finished" podID="7980021c-dde3-4c14-a7b6-fbcc947a183d" containerID="ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365" exitCode=1 Nov 24 06:49:31 crc kubenswrapper[4799]: I1124 06:49:31.430658 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvd7f" event={"ID":"7980021c-dde3-4c14-a7b6-fbcc947a183d","Type":"ContainerDied","Data":"ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365"} Nov 24 06:49:31 crc kubenswrapper[4799]: I1124 06:49:31.430757 4799 scope.go:117] "RemoveContainer" containerID="ed2993e8b921da2bb684c48300250cf3ddd1b308d455c4aa75304514a9700e6d" Nov 24 06:49:31 crc kubenswrapper[4799]: I1124 06:49:31.431518 4799 scope.go:117] "RemoveContainer" containerID="ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365" Nov 24 06:49:31 crc kubenswrapper[4799]: E1124 06:49:31.432783 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-zvd7f_openshift-multus(7980021c-dde3-4c14-a7b6-fbcc947a183d)\"" pod="openshift-multus/multus-zvd7f" podUID="7980021c-dde3-4c14-a7b6-fbcc947a183d" Nov 24 06:49:31 crc kubenswrapper[4799]: I1124 06:49:31.460851 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nvx7t" podStartSLOduration=95.460814063 podStartE2EDuration="1m35.460814063s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:15.389699867 +0000 UTC m=+101.045682371" watchObservedRunningTime="2025-11-24 06:49:31.460814063 +0000 UTC m=+117.116796567" Nov 24 06:49:32 crc kubenswrapper[4799]: I1124 06:49:32.437496 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvd7f_7980021c-dde3-4c14-a7b6-fbcc947a183d/kube-multus/1.log" Nov 24 06:49:32 crc kubenswrapper[4799]: I1124 06:49:32.628207 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:32 crc kubenswrapper[4799]: I1124 06:49:32.628254 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:32 crc kubenswrapper[4799]: I1124 06:49:32.628207 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:32 crc kubenswrapper[4799]: I1124 06:49:32.628798 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:32 crc kubenswrapper[4799]: E1124 06:49:32.629059 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:32 crc kubenswrapper[4799]: E1124 06:49:32.629250 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:32 crc kubenswrapper[4799]: E1124 06:49:32.629434 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:32 crc kubenswrapper[4799]: E1124 06:49:32.629603 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:34 crc kubenswrapper[4799]: I1124 06:49:34.627439 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:34 crc kubenswrapper[4799]: I1124 06:49:34.627468 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:34 crc kubenswrapper[4799]: I1124 06:49:34.627524 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:34 crc kubenswrapper[4799]: I1124 06:49:34.627489 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:34 crc kubenswrapper[4799]: E1124 06:49:34.627779 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:34 crc kubenswrapper[4799]: E1124 06:49:34.627938 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:34 crc kubenswrapper[4799]: E1124 06:49:34.628174 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:34 crc kubenswrapper[4799]: E1124 06:49:34.628256 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:34 crc kubenswrapper[4799]: I1124 06:49:34.629246 4799 scope.go:117] "RemoveContainer" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:49:34 crc kubenswrapper[4799]: E1124 06:49:34.629466 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b7nd7_openshift-ovn-kubernetes(5c34c957-e45b-4c65-8d6a-60e0a45b2f25)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" Nov 24 06:49:35 crc kubenswrapper[4799]: I1124 06:49:35.965153 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:35 crc kubenswrapper[4799]: E1124 06:49:35.965385 4799 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 06:49:35 crc kubenswrapper[4799]: E1124 06:49:35.968539 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:35 crc kubenswrapper[4799]: I1124 06:49:35.968591 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:35 crc kubenswrapper[4799]: I1124 06:49:35.968602 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:35 crc kubenswrapper[4799]: I1124 06:49:35.968633 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:35 crc kubenswrapper[4799]: E1124 06:49:35.969529 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:35 crc kubenswrapper[4799]: E1124 06:49:35.969715 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:35 crc kubenswrapper[4799]: E1124 06:49:35.969956 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:37 crc kubenswrapper[4799]: I1124 06:49:37.628196 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:37 crc kubenswrapper[4799]: I1124 06:49:37.628319 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:37 crc kubenswrapper[4799]: I1124 06:49:37.628218 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:37 crc kubenswrapper[4799]: E1124 06:49:37.628466 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:37 crc kubenswrapper[4799]: E1124 06:49:37.628643 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:37 crc kubenswrapper[4799]: I1124 06:49:37.628753 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:37 crc kubenswrapper[4799]: E1124 06:49:37.628917 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:37 crc kubenswrapper[4799]: E1124 06:49:37.629055 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:39 crc kubenswrapper[4799]: I1124 06:49:39.627603 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:39 crc kubenswrapper[4799]: I1124 06:49:39.627676 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:39 crc kubenswrapper[4799]: I1124 06:49:39.628268 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:39 crc kubenswrapper[4799]: I1124 06:49:39.628379 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:39 crc kubenswrapper[4799]: E1124 06:49:39.628781 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:39 crc kubenswrapper[4799]: E1124 06:49:39.629027 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:39 crc kubenswrapper[4799]: E1124 06:49:39.629177 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:39 crc kubenswrapper[4799]: E1124 06:49:39.629334 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:40 crc kubenswrapper[4799]: E1124 06:49:40.739781 4799 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 06:49:41 crc kubenswrapper[4799]: I1124 06:49:41.627478 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:41 crc kubenswrapper[4799]: I1124 06:49:41.627567 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:41 crc kubenswrapper[4799]: I1124 06:49:41.627478 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:41 crc kubenswrapper[4799]: I1124 06:49:41.627686 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:41 crc kubenswrapper[4799]: E1124 06:49:41.627699 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:41 crc kubenswrapper[4799]: E1124 06:49:41.628167 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:41 crc kubenswrapper[4799]: E1124 06:49:41.628295 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:41 crc kubenswrapper[4799]: E1124 06:49:41.628379 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:43 crc kubenswrapper[4799]: I1124 06:49:43.628274 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:43 crc kubenswrapper[4799]: I1124 06:49:43.628382 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:43 crc kubenswrapper[4799]: I1124 06:49:43.628311 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:43 crc kubenswrapper[4799]: E1124 06:49:43.628517 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:43 crc kubenswrapper[4799]: E1124 06:49:43.628731 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:43 crc kubenswrapper[4799]: I1124 06:49:43.628879 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:43 crc kubenswrapper[4799]: E1124 06:49:43.629270 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:43 crc kubenswrapper[4799]: I1124 06:49:43.629485 4799 scope.go:117] "RemoveContainer" containerID="ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365" Nov 24 06:49:43 crc kubenswrapper[4799]: E1124 06:49:43.629471 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:44 crc kubenswrapper[4799]: I1124 06:49:44.493397 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvd7f_7980021c-dde3-4c14-a7b6-fbcc947a183d/kube-multus/1.log" Nov 24 06:49:44 crc kubenswrapper[4799]: I1124 06:49:44.493981 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvd7f" event={"ID":"7980021c-dde3-4c14-a7b6-fbcc947a183d","Type":"ContainerStarted","Data":"726b153e9aa2f296175c87a8e127b3e7ae73ed29881843f9bb9b1220f6b95d95"} Nov 24 06:49:45 crc kubenswrapper[4799]: I1124 06:49:45.628002 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:45 crc kubenswrapper[4799]: I1124 06:49:45.628030 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:45 crc kubenswrapper[4799]: I1124 06:49:45.628003 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:45 crc kubenswrapper[4799]: I1124 06:49:45.628162 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:45 crc kubenswrapper[4799]: E1124 06:49:45.629499 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:45 crc kubenswrapper[4799]: E1124 06:49:45.629906 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:45 crc kubenswrapper[4799]: E1124 06:49:45.629988 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:45 crc kubenswrapper[4799]: E1124 06:49:45.630045 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:45 crc kubenswrapper[4799]: E1124 06:49:45.741070 4799 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 06:49:46 crc kubenswrapper[4799]: I1124 06:49:46.629688 4799 scope.go:117] "RemoveContainer" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:49:47 crc kubenswrapper[4799]: I1124 06:49:47.510835 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/3.log" Nov 24 06:49:47 crc kubenswrapper[4799]: I1124 06:49:47.515699 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerStarted","Data":"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2"} Nov 24 06:49:47 crc kubenswrapper[4799]: I1124 06:49:47.516571 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:49:47 crc kubenswrapper[4799]: I1124 06:49:47.555296 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ldbr4"] Nov 24 06:49:47 crc kubenswrapper[4799]: I1124 06:49:47.555524 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:47 crc kubenswrapper[4799]: E1124 06:49:47.555745 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:47 crc kubenswrapper[4799]: I1124 06:49:47.581624 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podStartSLOduration=111.581590169 podStartE2EDuration="1m51.581590169s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:47.578688941 +0000 UTC m=+133.234671455" watchObservedRunningTime="2025-11-24 06:49:47.581590169 +0000 UTC m=+133.237572673" Nov 24 06:49:47 crc kubenswrapper[4799]: I1124 06:49:47.627745 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:47 crc kubenswrapper[4799]: I1124 06:49:47.627814 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:47 crc kubenswrapper[4799]: I1124 06:49:47.627930 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:47 crc kubenswrapper[4799]: E1124 06:49:47.627958 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:47 crc kubenswrapper[4799]: E1124 06:49:47.628090 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:47 crc kubenswrapper[4799]: E1124 06:49:47.628315 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:49 crc kubenswrapper[4799]: I1124 06:49:49.628234 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:49 crc kubenswrapper[4799]: I1124 06:49:49.628234 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:49 crc kubenswrapper[4799]: E1124 06:49:49.628952 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ldbr4" podUID="f587ff07-479b-4c95-95cb-406faffcfb68" Nov 24 06:49:49 crc kubenswrapper[4799]: I1124 06:49:49.628436 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:49 crc kubenswrapper[4799]: I1124 06:49:49.628394 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:49 crc kubenswrapper[4799]: E1124 06:49:49.629138 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 06:49:49 crc kubenswrapper[4799]: E1124 06:49:49.629421 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 06:49:49 crc kubenswrapper[4799]: E1124 06:49:49.629531 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 06:49:51 crc kubenswrapper[4799]: I1124 06:49:51.627561 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:49:51 crc kubenswrapper[4799]: I1124 06:49:51.627607 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:49:51 crc kubenswrapper[4799]: I1124 06:49:51.627647 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:49:51 crc kubenswrapper[4799]: I1124 06:49:51.627582 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:49:51 crc kubenswrapper[4799]: I1124 06:49:51.630556 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 06:49:51 crc kubenswrapper[4799]: I1124 06:49:51.631714 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 06:49:51 crc kubenswrapper[4799]: I1124 06:49:51.631811 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 06:49:51 crc kubenswrapper[4799]: I1124 06:49:51.632370 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 06:49:51 crc kubenswrapper[4799]: I1124 06:49:51.633561 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 06:49:51 crc kubenswrapper[4799]: I1124 06:49:51.636168 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.866454 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.923137 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qv7f9"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.923777 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.924269 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.924492 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.928138 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.928352 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.928487 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.928491 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.928647 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.928668 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nqlvh"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.929643 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.932884 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rzh22"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.933337 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.934142 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.935789 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zzd4b"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.936675 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.936680 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.937736 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.937750 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rhhsg"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.940768 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-glxff"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.941405 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.941823 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-b28hc"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.942684 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.944129 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.944597 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.944894 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.945155 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.945422 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.945576 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.945693 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.945717 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.945955 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.945982 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.946468 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.946562 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.946927 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.947777 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.951819 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.952329 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.952482 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.952649 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.952669 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.952771 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.952972 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.953104 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.953205 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.953444 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.953595 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.955069 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.955100 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.957162 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.957446 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.959826 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.962202 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.965930 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.965968 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.966176 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.966277 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.966498 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.966650 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.966751 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.967096 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.967148 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.967226 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl"] Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.968767 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.970229 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.985214 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.986277 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 06:49:54 crc kubenswrapper[4799]: I1124 06:49:54.993004 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t6hw2"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.013601 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.014208 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.020133 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.021870 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.021879 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.022042 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.022155 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.022199 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.022957 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.023013 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.023206 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.023327 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.023444 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.023474 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.023556 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.023749 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.023778 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.023959 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.024066 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.028544 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-6j6bn"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.028936 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.029096 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-58mkm"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.029262 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6j6bn" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.029354 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.030700 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.031133 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5bjhn"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.031435 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gmz9"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.033826 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.033881 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.033963 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.034156 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.035310 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.036310 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.037562 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.038068 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.040266 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.040330 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.040282 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.040491 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.040743 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.040752 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.040782 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.041055 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.041133 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.041182 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.041198 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.041384 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.041798 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.041954 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.041977 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.042968 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.043369 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.043710 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.044542 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.053565 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.054067 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.054318 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.054569 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.054706 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.054916 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.054999 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.055077 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.055157 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.055236 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.056948 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.057549 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.058218 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.058407 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.058695 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.058912 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.061283 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-etcd-serving-ca\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.061329 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2ks5\" (UniqueName: \"kubernetes.io/projected/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-kube-api-access-v2ks5\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.061368 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-client-ca\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.061396 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.061426 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.061459 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12dc0898-d8d0-4b23-a8e2-3ed45987583b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fd7xr\" (UID: \"12dc0898-d8d0-4b23-a8e2-3ed45987583b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.061485 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a63db911-38e5-47a8-926a-5181ffa957bc-serving-cert\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.061513 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/facf7448-3c40-476b-bc7f-9526c5f4c3e2-serving-cert\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.061537 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed31a762-47f8-4efa-88f4-262e93cbdcea-serving-cert\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.061558 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.061562 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b717be8-b446-40c0-a5d9-1de2525168fd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-njqfl\" (UID: \"9b717be8-b446-40c0-a5d9-1de2525168fd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.070876 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fec9f210-0d28-463f-b3bf-4f27c6eee903-metrics-tls\") pod \"dns-operator-744455d44c-nqlvh\" (UID: \"fec9f210-0d28-463f-b3bf-4f27c6eee903\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.070996 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed31a762-47f8-4efa-88f4-262e93cbdcea-etcd-client\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.071032 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-config\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.071090 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.071220 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90dfb996-eeb4-4004-8879-86be3f6a9e66-serving-cert\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.071269 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-audit\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.071327 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-node-pullsecrets\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.071575 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg4jq\" (UniqueName: \"kubernetes.io/projected/facf7448-3c40-476b-bc7f-9526c5f4c3e2-kube-api-access-kg4jq\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.071613 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90dfb996-eeb4-4004-8879-86be3f6a9e66-service-ca-bundle\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.071743 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36f481dd-d3ab-422a-8f26-69fcccb51410-serving-cert\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.071777 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88h7l\" (UniqueName: \"kubernetes.io/projected/bfae07f4-0529-426f-aaac-d82964808045-kube-api-access-88h7l\") pod \"openshift-config-operator-7777fb866f-glxff\" (UID: \"bfae07f4-0529-426f-aaac-d82964808045\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.071806 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6hqg\" (UniqueName: \"kubernetes.io/projected/9b717be8-b446-40c0-a5d9-1de2525168fd-kube-api-access-j6hqg\") pod \"cluster-samples-operator-665b6dd947-njqfl\" (UID: \"9b717be8-b446-40c0-a5d9-1de2525168fd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.072008 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gttgg\" (UniqueName: \"kubernetes.io/projected/36f481dd-d3ab-422a-8f26-69fcccb51410-kube-api-access-gttgg\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.072041 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a63db911-38e5-47a8-926a-5181ffa957bc-trusted-ca\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.072088 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-serving-cert\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.072280 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0119fbf1-7180-4670-a389-adcf32d1a736-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwzs\" (UID: \"0119fbf1-7180-4670-a389-adcf32d1a736\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.072314 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-encryption-config\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.072507 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-audit-dir\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.072875 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgx2s\" (UniqueName: \"kubernetes.io/projected/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-kube-api-access-bgx2s\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.072917 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12dc0898-d8d0-4b23-a8e2-3ed45987583b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fd7xr\" (UID: \"12dc0898-d8d0-4b23-a8e2-3ed45987583b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.072954 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfae07f4-0529-426f-aaac-d82964808045-serving-cert\") pod \"openshift-config-operator-7777fb866f-glxff\" (UID: \"bfae07f4-0529-426f-aaac-d82964808045\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.072975 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ed31a762-47f8-4efa-88f4-262e93cbdcea-etcd-service-ca\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073003 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmzkl\" (UniqueName: \"kubernetes.io/projected/a63db911-38e5-47a8-926a-5181ffa957bc-kube-api-access-qmzkl\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073032 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073221 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffmvj\" (UniqueName: \"kubernetes.io/projected/ed31a762-47f8-4efa-88f4-262e93cbdcea-kube-api-access-ffmvj\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073244 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvkqx\" (UniqueName: \"kubernetes.io/projected/0119fbf1-7180-4670-a389-adcf32d1a736-kube-api-access-tvkqx\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwzs\" (UID: \"0119fbf1-7180-4670-a389-adcf32d1a736\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073270 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90dfb996-eeb4-4004-8879-86be3f6a9e66-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073297 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-etcd-client\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073377 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96z77\" (UniqueName: \"kubernetes.io/projected/12dc0898-d8d0-4b23-a8e2-3ed45987583b-kube-api-access-96z77\") pod \"openshift-controller-manager-operator-756b6f6bc6-fd7xr\" (UID: \"12dc0898-d8d0-4b23-a8e2-3ed45987583b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073451 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0119fbf1-7180-4670-a389-adcf32d1a736-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwzs\" (UID: \"0119fbf1-7180-4670-a389-adcf32d1a736\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073522 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63db911-38e5-47a8-926a-5181ffa957bc-config\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073653 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjkzw\" (UniqueName: \"kubernetes.io/projected/90dfb996-eeb4-4004-8879-86be3f6a9e66-kube-api-access-gjkzw\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073684 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073757 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-image-import-ca\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073791 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-config\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.073818 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-client-ca\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.083438 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkhk9\" (UniqueName: \"kubernetes.io/projected/fec9f210-0d28-463f-b3bf-4f27c6eee903-kube-api-access-xkhk9\") pod \"dns-operator-744455d44c-nqlvh\" (UID: \"fec9f210-0d28-463f-b3bf-4f27c6eee903\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.083899 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.084542 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bfae07f4-0529-426f-aaac-d82964808045-available-featuregates\") pod \"openshift-config-operator-7777fb866f-glxff\" (UID: \"bfae07f4-0529-426f-aaac-d82964808045\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.084582 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ed31a762-47f8-4efa-88f4-262e93cbdcea-etcd-ca\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.084658 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-config\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.085120 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed31a762-47f8-4efa-88f4-262e93cbdcea-config\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.092347 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.092909 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.094660 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.095517 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90dfb996-eeb4-4004-8879-86be3f6a9e66-config\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.095788 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.097600 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.100984 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-cn7ch"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.102570 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.102837 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.102982 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.103933 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.104369 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.104934 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.105481 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.106217 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.108051 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.111741 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.113819 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.114859 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.115539 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.116536 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rzh22"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.117408 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.118148 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.119207 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qv7f9"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.120124 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.121109 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zzd4b"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.121129 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.122337 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-g4h5t"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.123139 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.123897 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.125142 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.125958 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.127033 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-64b4w"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.127625 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.127669 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.129349 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.129813 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.131237 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.131682 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.131883 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcvtw"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.132799 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.134741 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.135673 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.137581 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.138322 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.138788 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.138902 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.139152 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.139648 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-bbm7w"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.140206 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.146704 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.157920 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-glxff"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.160155 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.169616 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.171943 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.174666 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nqlvh"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.176354 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.180205 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5bjhn"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.182702 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.186655 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.186719 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.189659 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.191118 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2zg6b"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.192321 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.192657 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.194535 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rhhsg"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.196553 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-27fxx"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197312 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfae07f4-0529-426f-aaac-d82964808045-serving-cert\") pod \"openshift-config-operator-7777fb866f-glxff\" (UID: \"bfae07f4-0529-426f-aaac-d82964808045\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197359 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ed31a762-47f8-4efa-88f4-262e93cbdcea-etcd-service-ca\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197385 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmzkl\" (UniqueName: \"kubernetes.io/projected/a63db911-38e5-47a8-926a-5181ffa957bc-kube-api-access-qmzkl\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197412 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197434 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1317991a-3615-4dc5-9cbe-232e04180a7c-metrics-certs\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197460 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffmvj\" (UniqueName: \"kubernetes.io/projected/ed31a762-47f8-4efa-88f4-262e93cbdcea-kube-api-access-ffmvj\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197477 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvkqx\" (UniqueName: \"kubernetes.io/projected/0119fbf1-7180-4670-a389-adcf32d1a736-kube-api-access-tvkqx\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwzs\" (UID: \"0119fbf1-7180-4670-a389-adcf32d1a736\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197497 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90dfb996-eeb4-4004-8879-86be3f6a9e66-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197517 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-etcd-client\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197535 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96z77\" (UniqueName: \"kubernetes.io/projected/12dc0898-d8d0-4b23-a8e2-3ed45987583b-kube-api-access-96z77\") pod \"openshift-controller-manager-operator-756b6f6bc6-fd7xr\" (UID: \"12dc0898-d8d0-4b23-a8e2-3ed45987583b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197727 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0119fbf1-7180-4670-a389-adcf32d1a736-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwzs\" (UID: \"0119fbf1-7180-4670-a389-adcf32d1a736\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197782 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-27fxx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197811 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63db911-38e5-47a8-926a-5181ffa957bc-config\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197838 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjkzw\" (UniqueName: \"kubernetes.io/projected/90dfb996-eeb4-4004-8879-86be3f6a9e66-kube-api-access-gjkzw\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197898 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197930 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-oauth-config\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197960 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-image-import-ca\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.197982 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pqdc\" (UniqueName: \"kubernetes.io/projected/336a419e-e807-4859-b624-2ed06a9a8665-kube-api-access-4pqdc\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198017 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-config\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198058 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-service-ca\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198091 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-oauth-serving-cert\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198117 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1317991a-3615-4dc5-9cbe-232e04180a7c-default-certificate\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198152 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-client-ca\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198174 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkhk9\" (UniqueName: \"kubernetes.io/projected/fec9f210-0d28-463f-b3bf-4f27c6eee903-kube-api-access-xkhk9\") pod \"dns-operator-744455d44c-nqlvh\" (UID: \"fec9f210-0d28-463f-b3bf-4f27c6eee903\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198192 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jqgz\" (UniqueName: \"kubernetes.io/projected/1317991a-3615-4dc5-9cbe-232e04180a7c-kube-api-access-7jqgz\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198216 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bfae07f4-0529-426f-aaac-d82964808045-available-featuregates\") pod \"openshift-config-operator-7777fb866f-glxff\" (UID: \"bfae07f4-0529-426f-aaac-d82964808045\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198241 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ed31a762-47f8-4efa-88f4-262e93cbdcea-etcd-ca\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198263 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-config\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198298 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed31a762-47f8-4efa-88f4-262e93cbdcea-config\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198321 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-trusted-ca-bundle\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90dfb996-eeb4-4004-8879-86be3f6a9e66-config\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198364 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tvjv\" (UniqueName: \"kubernetes.io/projected/86f640cf-7c54-4b44-a52f-21378c535d5a-kube-api-access-2tvjv\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198386 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2ks5\" (UniqueName: \"kubernetes.io/projected/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-kube-api-access-v2ks5\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198404 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-etcd-serving-ca\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198407 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ed31a762-47f8-4efa-88f4-262e93cbdcea-etcd-service-ca\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198432 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-client-ca\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198451 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198477 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/86f640cf-7c54-4b44-a52f-21378c535d5a-images\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198501 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198522 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86f640cf-7c54-4b44-a52f-21378c535d5a-config\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198521 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-b28hc"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198547 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12dc0898-d8d0-4b23-a8e2-3ed45987583b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fd7xr\" (UID: \"12dc0898-d8d0-4b23-a8e2-3ed45987583b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198595 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fec9f210-0d28-463f-b3bf-4f27c6eee903-metrics-tls\") pod \"dns-operator-744455d44c-nqlvh\" (UID: \"fec9f210-0d28-463f-b3bf-4f27c6eee903\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198621 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a63db911-38e5-47a8-926a-5181ffa957bc-serving-cert\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198639 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/facf7448-3c40-476b-bc7f-9526c5f4c3e2-serving-cert\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198713 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed31a762-47f8-4efa-88f4-262e93cbdcea-serving-cert\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198781 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b717be8-b446-40c0-a5d9-1de2525168fd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-njqfl\" (UID: \"9b717be8-b446-40c0-a5d9-1de2525168fd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198800 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-serving-cert\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198832 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed31a762-47f8-4efa-88f4-262e93cbdcea-etcd-client\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198865 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-config\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198886 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/86f640cf-7c54-4b44-a52f-21378c535d5a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198909 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198924 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-console-config\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.198997 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-audit\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199035 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90dfb996-eeb4-4004-8879-86be3f6a9e66-serving-cert\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199057 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-node-pullsecrets\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199075 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1317991a-3615-4dc5-9cbe-232e04180a7c-service-ca-bundle\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199093 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg4jq\" (UniqueName: \"kubernetes.io/projected/facf7448-3c40-476b-bc7f-9526c5f4c3e2-kube-api-access-kg4jq\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199110 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90dfb996-eeb4-4004-8879-86be3f6a9e66-service-ca-bundle\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199126 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36f481dd-d3ab-422a-8f26-69fcccb51410-serving-cert\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199148 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88h7l\" (UniqueName: \"kubernetes.io/projected/bfae07f4-0529-426f-aaac-d82964808045-kube-api-access-88h7l\") pod \"openshift-config-operator-7777fb866f-glxff\" (UID: \"bfae07f4-0529-426f-aaac-d82964808045\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199180 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6hqg\" (UniqueName: \"kubernetes.io/projected/9b717be8-b446-40c0-a5d9-1de2525168fd-kube-api-access-j6hqg\") pod \"cluster-samples-operator-665b6dd947-njqfl\" (UID: \"9b717be8-b446-40c0-a5d9-1de2525168fd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199207 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a63db911-38e5-47a8-926a-5181ffa957bc-trusted-ca\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199213 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199244 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-serving-cert\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199270 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gttgg\" (UniqueName: \"kubernetes.io/projected/36f481dd-d3ab-422a-8f26-69fcccb51410-kube-api-access-gttgg\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199301 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0119fbf1-7180-4670-a389-adcf32d1a736-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwzs\" (UID: \"0119fbf1-7180-4670-a389-adcf32d1a736\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199321 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-encryption-config\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199338 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-audit-dir\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199358 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgx2s\" (UniqueName: \"kubernetes.io/projected/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-kube-api-access-bgx2s\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199384 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12dc0898-d8d0-4b23-a8e2-3ed45987583b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fd7xr\" (UID: \"12dc0898-d8d0-4b23-a8e2-3ed45987583b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199411 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1317991a-3615-4dc5-9cbe-232e04180a7c-stats-auth\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.199898 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90dfb996-eeb4-4004-8879-86be3f6a9e66-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.200149 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0119fbf1-7180-4670-a389-adcf32d1a736-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwzs\" (UID: \"0119fbf1-7180-4670-a389-adcf32d1a736\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.200316 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63db911-38e5-47a8-926a-5181ffa957bc-config\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.200464 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.200326 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bfae07f4-0529-426f-aaac-d82964808045-available-featuregates\") pod \"openshift-config-operator-7777fb866f-glxff\" (UID: \"bfae07f4-0529-426f-aaac-d82964808045\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.200608 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed31a762-47f8-4efa-88f4-262e93cbdcea-config\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.200636 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-config\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.200698 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90dfb996-eeb4-4004-8879-86be3f6a9e66-config\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.200741 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-client-ca\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.201750 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-etcd-serving-ca\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.201862 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-config\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.201991 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.202759 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-audit-dir\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.202837 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-config\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.202971 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a63db911-38e5-47a8-926a-5181ffa957bc-trusted-ca\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.203470 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-image-import-ca\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.203555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-client-ca\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.203718 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.203774 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t6hw2"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.203874 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.204116 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcvtw"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.204675 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12dc0898-d8d0-4b23-a8e2-3ed45987583b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fd7xr\" (UID: \"12dc0898-d8d0-4b23-a8e2-3ed45987583b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.205318 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-node-pullsecrets\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.206166 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ed31a762-47f8-4efa-88f4-262e93cbdcea-etcd-ca\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.206527 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.206712 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.206760 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-g4h5t"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.206973 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.207025 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90dfb996-eeb4-4004-8879-86be3f6a9e66-service-ca-bundle\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.207648 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.208322 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-audit\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.208476 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-dv8qs"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.208760 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-serving-cert\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.209285 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0119fbf1-7180-4670-a389-adcf32d1a736-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwzs\" (UID: \"0119fbf1-7180-4670-a389-adcf32d1a736\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.209516 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed31a762-47f8-4efa-88f4-262e93cbdcea-etcd-client\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.209677 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/facf7448-3c40-476b-bc7f-9526c5f4c3e2-serving-cert\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.210056 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6j6bn"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.210169 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-dv8qs" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.210734 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfae07f4-0529-426f-aaac-d82964808045-serving-cert\") pod \"openshift-config-operator-7777fb866f-glxff\" (UID: \"bfae07f4-0529-426f-aaac-d82964808045\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.210832 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed31a762-47f8-4efa-88f4-262e93cbdcea-serving-cert\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.210897 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-64b4w"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.210998 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b717be8-b446-40c0-a5d9-1de2525168fd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-njqfl\" (UID: \"9b717be8-b446-40c0-a5d9-1de2525168fd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.211464 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-encryption-config\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.211632 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12dc0898-d8d0-4b23-a8e2-3ed45987583b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fd7xr\" (UID: \"12dc0898-d8d0-4b23-a8e2-3ed45987583b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.212014 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fec9f210-0d28-463f-b3bf-4f27c6eee903-metrics-tls\") pod \"dns-operator-744455d44c-nqlvh\" (UID: \"fec9f210-0d28-463f-b3bf-4f27c6eee903\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.212210 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90dfb996-eeb4-4004-8879-86be3f6a9e66-serving-cert\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.212225 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.212438 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-etcd-client\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.213936 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-w4d6j"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.214372 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.214624 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36f481dd-d3ab-422a-8f26-69fcccb51410-serving-cert\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.215174 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2zg6b"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.215200 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-bbm7w"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.215394 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-w4d6j" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.215424 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a63db911-38e5-47a8-926a-5181ffa957bc-serving-cert\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.216546 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.218083 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.219379 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.220505 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.222000 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gmz9"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.223152 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-58mkm"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.224249 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-27fxx"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.225399 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-w4d6j"] Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.227796 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.246371 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.267579 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.288558 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.300615 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-trusted-ca-bundle\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.300666 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tvjv\" (UniqueName: \"kubernetes.io/projected/86f640cf-7c54-4b44-a52f-21378c535d5a-kube-api-access-2tvjv\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.300707 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/86f640cf-7c54-4b44-a52f-21378c535d5a-images\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.300736 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86f640cf-7c54-4b44-a52f-21378c535d5a-config\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.300778 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/86f640cf-7c54-4b44-a52f-21378c535d5a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.300802 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-serving-cert\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.300831 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-console-config\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.300903 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1317991a-3615-4dc5-9cbe-232e04180a7c-service-ca-bundle\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.301004 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1317991a-3615-4dc5-9cbe-232e04180a7c-stats-auth\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.301062 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1317991a-3615-4dc5-9cbe-232e04180a7c-metrics-certs\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.301151 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-oauth-config\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.301177 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pqdc\" (UniqueName: \"kubernetes.io/projected/336a419e-e807-4859-b624-2ed06a9a8665-kube-api-access-4pqdc\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.301200 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-service-ca\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.301232 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-oauth-serving-cert\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.301254 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1317991a-3615-4dc5-9cbe-232e04180a7c-default-certificate\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.301275 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jqgz\" (UniqueName: \"kubernetes.io/projected/1317991a-3615-4dc5-9cbe-232e04180a7c-kube-api-access-7jqgz\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.301964 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/86f640cf-7c54-4b44-a52f-21378c535d5a-images\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.302052 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86f640cf-7c54-4b44-a52f-21378c535d5a-config\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.302825 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-oauth-serving-cert\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.303207 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-console-config\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.303680 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-service-ca\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.303909 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-trusted-ca-bundle\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.304620 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/86f640cf-7c54-4b44-a52f-21378c535d5a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.305927 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-serving-cert\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.307019 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.317527 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-oauth-config\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.328491 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.346622 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.369497 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.387466 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.406328 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.447818 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.467961 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.488537 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.506709 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.527381 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.548090 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.568068 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.577664 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1317991a-3615-4dc5-9cbe-232e04180a7c-default-certificate\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.587771 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.596900 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1317991a-3615-4dc5-9cbe-232e04180a7c-metrics-certs\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.607107 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.614295 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1317991a-3615-4dc5-9cbe-232e04180a7c-service-ca-bundle\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.627781 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.638055 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1317991a-3615-4dc5-9cbe-232e04180a7c-stats-auth\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.647842 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.667792 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.687678 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.707229 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.727426 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.747808 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.767660 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.787680 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.807664 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.827331 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.848444 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.867889 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.888210 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.906900 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.928202 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.947358 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.968282 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 06:49:55 crc kubenswrapper[4799]: I1124 06:49:55.988626 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.007655 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.027590 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.047081 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.067097 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.088659 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.127575 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.145537 4799 request.go:700] Waited for 1.019297769s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serviceaccount-dockercfg-rq7zk&limit=500&resourceVersion=0 Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.148221 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.167757 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.187753 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.207940 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.228930 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.249486 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.266762 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.287675 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.307008 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.326324 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.351741 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.366741 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.386233 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.407738 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.427415 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.447112 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.480398 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.486707 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.506654 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.528027 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.546552 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.568093 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.587225 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.607700 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.627194 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.646606 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.667016 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.687127 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.706557 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.728149 4799 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.747221 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.767494 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.811933 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmzkl\" (UniqueName: \"kubernetes.io/projected/a63db911-38e5-47a8-926a-5181ffa957bc-kube-api-access-qmzkl\") pod \"console-operator-58897d9998-zzd4b\" (UID: \"a63db911-38e5-47a8-926a-5181ffa957bc\") " pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.829224 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvkqx\" (UniqueName: \"kubernetes.io/projected/0119fbf1-7180-4670-a389-adcf32d1a736-kube-api-access-tvkqx\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwzs\" (UID: \"0119fbf1-7180-4670-a389-adcf32d1a736\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.854459 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffmvj\" (UniqueName: \"kubernetes.io/projected/ed31a762-47f8-4efa-88f4-262e93cbdcea-kube-api-access-ffmvj\") pod \"etcd-operator-b45778765-qv7f9\" (UID: \"ed31a762-47f8-4efa-88f4-262e93cbdcea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.867661 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.867881 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96z77\" (UniqueName: \"kubernetes.io/projected/12dc0898-d8d0-4b23-a8e2-3ed45987583b-kube-api-access-96z77\") pod \"openshift-controller-manager-operator-756b6f6bc6-fd7xr\" (UID: \"12dc0898-d8d0-4b23-a8e2-3ed45987583b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.887616 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.908000 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.908197 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.920904 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.928167 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.973650 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:56 crc kubenswrapper[4799]: I1124 06:49:56.996082 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkhk9\" (UniqueName: \"kubernetes.io/projected/fec9f210-0d28-463f-b3bf-4f27c6eee903-kube-api-access-xkhk9\") pod \"dns-operator-744455d44c-nqlvh\" (UID: \"fec9f210-0d28-463f-b3bf-4f27c6eee903\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.016380 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjkzw\" (UniqueName: \"kubernetes.io/projected/90dfb996-eeb4-4004-8879-86be3f6a9e66-kube-api-access-gjkzw\") pod \"authentication-operator-69f744f599-rhhsg\" (UID: \"90dfb996-eeb4-4004-8879-86be3f6a9e66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.044554 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6hqg\" (UniqueName: \"kubernetes.io/projected/9b717be8-b446-40c0-a5d9-1de2525168fd-kube-api-access-j6hqg\") pod \"cluster-samples-operator-665b6dd947-njqfl\" (UID: \"9b717be8-b446-40c0-a5d9-1de2525168fd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.054497 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.058572 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2ks5\" (UniqueName: \"kubernetes.io/projected/50f64a8f-2611-4be4-b8e4-08eba7fa0e69-kube-api-access-v2ks5\") pod \"cluster-image-registry-operator-dc59b4c8b-hqzcz\" (UID: \"50f64a8f-2611-4be4-b8e4-08eba7fa0e69\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.086904 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg4jq\" (UniqueName: \"kubernetes.io/projected/facf7448-3c40-476b-bc7f-9526c5f4c3e2-kube-api-access-kg4jq\") pod \"route-controller-manager-6576b87f9c-77hvx\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.100555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgx2s\" (UniqueName: \"kubernetes.io/projected/1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f-kube-api-access-bgx2s\") pod \"apiserver-76f77b778f-b28hc\" (UID: \"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f\") " pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.114740 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gttgg\" (UniqueName: \"kubernetes.io/projected/36f481dd-d3ab-422a-8f26-69fcccb51410-kube-api-access-gttgg\") pod \"controller-manager-879f6c89f-rzh22\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.124264 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.129659 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.134233 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88h7l\" (UniqueName: \"kubernetes.io/projected/bfae07f4-0529-426f-aaac-d82964808045-kube-api-access-88h7l\") pod \"openshift-config-operator-7777fb866f-glxff\" (UID: \"bfae07f4-0529-426f-aaac-d82964808045\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.147277 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.147665 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.165069 4799 request.go:700] Waited for 1.954475485s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-tls&limit=500&resourceVersion=0 Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.168506 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.187978 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.196164 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.208060 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.219660 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs"] Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.227429 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.234339 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.242715 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.247712 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.263432 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.272696 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jqgz\" (UniqueName: \"kubernetes.io/projected/1317991a-3615-4dc5-9cbe-232e04180a7c-kube-api-access-7jqgz\") pod \"router-default-5444994796-cn7ch\" (UID: \"1317991a-3615-4dc5-9cbe-232e04180a7c\") " pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.277583 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.283761 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.287637 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pqdc\" (UniqueName: \"kubernetes.io/projected/336a419e-e807-4859-b624-2ed06a9a8665-kube-api-access-4pqdc\") pod \"console-f9d7485db-58mkm\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.305290 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tvjv\" (UniqueName: \"kubernetes.io/projected/86f640cf-7c54-4b44-a52f-21378c535d5a-kube-api-access-2tvjv\") pod \"machine-api-operator-5694c8668f-t6hw2\" (UID: \"86f640cf-7c54-4b44-a52f-21378c535d5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.312835 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr"] Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.348922 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.348983 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/892c2271-ba94-41b9-b5ed-04d45c2b1194-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349005 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-registry-certificates\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349091 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8kb6\" (UniqueName: \"kubernetes.io/projected/4708bfd8-5610-47ff-a842-d2bc1c6e05bc-kube-api-access-g8kb6\") pod \"machine-config-controller-84d6567774-n2sth\" (UID: \"4708bfd8-5610-47ff-a842-d2bc1c6e05bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349119 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349140 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349165 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/892c2271-ba94-41b9-b5ed-04d45c2b1194-images\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349184 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4708bfd8-5610-47ff-a842-d2bc1c6e05bc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-n2sth\" (UID: \"4708bfd8-5610-47ff-a842-d2bc1c6e05bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349200 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17a805ea-4e7f-4204-9712-8be2c673363d-metrics-tls\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349239 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nf9tt\" (UID: \"c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349263 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf-config\") pod \"kube-apiserver-operator-766d6c64bb-nf9tt\" (UID: \"c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349280 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22faf60d-16a2-49ae-9eb5-5f808013081e-serving-cert\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349296 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b7761980-8663-4dda-924b-2fb787fcdac8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349317 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349370 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47dvt\" (UniqueName: \"kubernetes.io/projected/17a805ea-4e7f-4204-9712-8be2c673363d-kube-api-access-47dvt\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349422 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/22faf60d-16a2-49ae-9eb5-5f808013081e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349439 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349455 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17a805ea-4e7f-4204-9712-8be2c673363d-trusted-ca\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349474 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkpjn\" (UniqueName: \"kubernetes.io/projected/fde84f5d-c045-4a70-9d7d-9f4ca71d4351-kube-api-access-gkpjn\") pod \"control-plane-machine-set-operator-78cbb6b69f-knplv\" (UID: \"fde84f5d-c045-4a70-9d7d-9f4ca71d4351\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-policies\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349546 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/31d8b755-9373-401c-8663-dba43a056750-machine-approver-tls\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349564 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/22faf60d-16a2-49ae-9eb5-5f808013081e-audit-policies\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349580 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmbs6\" (UniqueName: \"kubernetes.io/projected/d6a12d20-1525-4cf5-8e5c-da7184c42581-kube-api-access-zmbs6\") pod \"downloads-7954f5f757-6j6bn\" (UID: \"d6a12d20-1525-4cf5-8e5c-da7184c42581\") " pod="openshift-console/downloads-7954f5f757-6j6bn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349604 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349630 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-registry-tls\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349659 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22faf60d-16a2-49ae-9eb5-5f808013081e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349695 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/17a805ea-4e7f-4204-9712-8be2c673363d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349722 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349742 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fde84f5d-c045-4a70-9d7d-9f4ca71d4351-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-knplv\" (UID: \"fde84f5d-c045-4a70-9d7d-9f4ca71d4351\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349760 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349777 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0625b922-f42b-4378-ba25-1ee3fc85cce6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2npgz\" (UID: \"0625b922-f42b-4378-ba25-1ee3fc85cce6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349792 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-trusted-ca\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349867 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6pv6\" (UniqueName: \"kubernetes.io/projected/61518cfd-2e9e-4d2f-95ed-612afa636cef-kube-api-access-r6pv6\") pod \"multus-admission-controller-857f4d67dd-g4h5t\" (UID: \"61518cfd-2e9e-4d2f-95ed-612afa636cef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349897 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/22faf60d-16a2-49ae-9eb5-5f808013081e-audit-dir\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349914 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdt74\" (UniqueName: \"kubernetes.io/projected/31d8b755-9373-401c-8663-dba43a056750-kube-api-access-bdt74\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349943 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-dir\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349960 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b755-9373-401c-8663-dba43a056750-auth-proxy-config\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.349983 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtzdz\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-kube-api-access-gtzdz\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350018 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-bound-sa-token\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350036 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350053 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0625b922-f42b-4378-ba25-1ee3fc85cce6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2npgz\" (UID: \"0625b922-f42b-4378-ba25-1ee3fc85cce6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350078 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b7761980-8663-4dda-924b-2fb787fcdac8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350110 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4708bfd8-5610-47ff-a842-d2bc1c6e05bc-proxy-tls\") pod \"machine-config-controller-84d6567774-n2sth\" (UID: \"4708bfd8-5610-47ff-a842-d2bc1c6e05bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350192 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/892c2271-ba94-41b9-b5ed-04d45c2b1194-proxy-tls\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350412 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350443 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgwr6\" (UniqueName: \"kubernetes.io/projected/892c2271-ba94-41b9-b5ed-04d45c2b1194-kube-api-access-mgwr6\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350478 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6bgh\" (UniqueName: \"kubernetes.io/projected/7dc41818-1f5e-4472-a726-f1322a1f462b-kube-api-access-l6bgh\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350530 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f394c75-c5d2-4b2a-8845-d2337bc1735e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jf7t\" (UID: \"7f394c75-c5d2-4b2a-8845-d2337bc1735e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350568 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nf9tt\" (UID: \"c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350588 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw9sc\" (UniqueName: \"kubernetes.io/projected/22faf60d-16a2-49ae-9eb5-5f808013081e-kube-api-access-lw9sc\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350653 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f394c75-c5d2-4b2a-8845-d2337bc1735e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jf7t\" (UID: \"7f394c75-c5d2-4b2a-8845-d2337bc1735e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350669 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/22faf60d-16a2-49ae-9eb5-5f808013081e-etcd-client\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350691 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/22faf60d-16a2-49ae-9eb5-5f808013081e-encryption-config\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350711 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0625b922-f42b-4378-ba25-1ee3fc85cce6-config\") pod \"kube-controller-manager-operator-78b949d7b-2npgz\" (UID: \"0625b922-f42b-4378-ba25-1ee3fc85cce6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350729 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350748 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b9m8\" (UniqueName: \"kubernetes.io/projected/5e8415c7-91a0-4d83-9627-d7e394393334-kube-api-access-6b9m8\") pod \"migrator-59844c95c7-79kln\" (UID: \"5e8415c7-91a0-4d83-9627-d7e394393334\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350775 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61518cfd-2e9e-4d2f-95ed-612afa636cef-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-g4h5t\" (UID: \"61518cfd-2e9e-4d2f-95ed-612afa636cef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350802 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f394c75-c5d2-4b2a-8845-d2337bc1735e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jf7t\" (UID: \"7f394c75-c5d2-4b2a-8845-d2337bc1735e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350839 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.350871 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31d8b755-9373-401c-8663-dba43a056750-config\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: E1124 06:49:57.353034 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:57.853014225 +0000 UTC m=+143.508996699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.356213 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.398898 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.427471 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zzd4b"] Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.451489 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:57 crc kubenswrapper[4799]: E1124 06:49:57.451894 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:57.951835563 +0000 UTC m=+143.607818037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.452139 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.452239 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.452321 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/892c2271-ba94-41b9-b5ed-04d45c2b1194-images\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.452410 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4708bfd8-5610-47ff-a842-d2bc1c6e05bc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-n2sth\" (UID: \"4708bfd8-5610-47ff-a842-d2bc1c6e05bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.452486 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7edcb436-bf5b-4afd-9603-4a8131b2be16-cert\") pod \"ingress-canary-27fxx\" (UID: \"7edcb436-bf5b-4afd-9603-4a8131b2be16\") " pod="openshift-ingress-canary/ingress-canary-27fxx" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.452553 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwdqk\" (UniqueName: \"kubernetes.io/projected/e116addf-dac0-42aa-9c0e-82383b214399-kube-api-access-mwdqk\") pod \"collect-profiles-29399445-2gfhd\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.452653 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nf9tt\" (UID: \"c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.452730 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf-config\") pod \"kube-apiserver-operator-766d6c64bb-nf9tt\" (UID: \"c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.452803 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22faf60d-16a2-49ae-9eb5-5f808013081e-serving-cert\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.453012 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17a805ea-4e7f-4204-9712-8be2c673363d-metrics-tls\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.453086 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzrrh\" (UniqueName: \"kubernetes.io/projected/3b02c876-c14b-4422-beaf-9ea3ab910f90-kube-api-access-xzrrh\") pod \"marketplace-operator-79b997595-pcvtw\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.453158 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b7761980-8663-4dda-924b-2fb787fcdac8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.453224 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.453350 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/baecc37e-a64f-449f-bde9-f4ff3a66e628-signing-key\") pod \"service-ca-9c57cc56f-bbm7w\" (UID: \"baecc37e-a64f-449f-bde9-f4ff3a66e628\") " pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.453431 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-mountpoint-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.453554 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4708bfd8-5610-47ff-a842-d2bc1c6e05bc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-n2sth\" (UID: \"4708bfd8-5610-47ff-a842-d2bc1c6e05bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.454142 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.454688 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/892c2271-ba94-41b9-b5ed-04d45c2b1194-images\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.456268 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b7761980-8663-4dda-924b-2fb787fcdac8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.456817 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf-config\") pod \"kube-apiserver-operator-766d6c64bb-nf9tt\" (UID: \"c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.466300 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-csi-data-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.466981 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47dvt\" (UniqueName: \"kubernetes.io/projected/17a805ea-4e7f-4204-9712-8be2c673363d-kube-api-access-47dvt\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467010 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f-node-bootstrap-token\") pod \"machine-config-server-dv8qs\" (UID: \"5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f\") " pod="openshift-machine-config-operator/machine-config-server-dv8qs" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467026 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f-certs\") pod \"machine-config-server-dv8qs\" (UID: \"5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f\") " pod="openshift-machine-config-operator/machine-config-server-dv8qs" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467049 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57pv7\" (UniqueName: \"kubernetes.io/projected/2927bbfb-073e-41b2-bb65-c6c75a54560b-kube-api-access-57pv7\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467119 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22faf60d-16a2-49ae-9eb5-5f808013081e-serving-cert\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.466505 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17a805ea-4e7f-4204-9712-8be2c673363d-metrics-tls\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467072 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7aede9e9-6b07-4959-8097-982b17d5628d-srv-cert\") pod \"catalog-operator-68c6474976-6d9nh\" (UID: \"7aede9e9-6b07-4959-8097-982b17d5628d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467274 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g8z8\" (UniqueName: \"kubernetes.io/projected/66a7f09b-b285-4635-8666-b8caaa05b1c7-kube-api-access-2g8z8\") pod \"dns-default-w4d6j\" (UID: \"66a7f09b-b285-4635-8666-b8caaa05b1c7\") " pod="openshift-dns/dns-default-w4d6j" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467309 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e116addf-dac0-42aa-9c0e-82383b214399-config-volume\") pod \"collect-profiles-29399445-2gfhd\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/22faf60d-16a2-49ae-9eb5-5f808013081e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467623 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467659 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17a805ea-4e7f-4204-9712-8be2c673363d-trusted-ca\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467690 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkpjn\" (UniqueName: \"kubernetes.io/projected/fde84f5d-c045-4a70-9d7d-9f4ca71d4351-kube-api-access-gkpjn\") pod \"control-plane-machine-set-operator-78cbb6b69f-knplv\" (UID: \"fde84f5d-c045-4a70-9d7d-9f4ca71d4351\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467739 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-policies\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467763 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/31d8b755-9373-401c-8663-dba43a056750-machine-approver-tls\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467787 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/22faf60d-16a2-49ae-9eb5-5f808013081e-audit-policies\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467805 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmbs6\" (UniqueName: \"kubernetes.io/projected/d6a12d20-1525-4cf5-8e5c-da7184c42581-kube-api-access-zmbs6\") pod \"downloads-7954f5f757-6j6bn\" (UID: \"d6a12d20-1525-4cf5-8e5c-da7184c42581\") " pod="openshift-console/downloads-7954f5f757-6j6bn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467879 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467903 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d5br\" (UniqueName: \"kubernetes.io/projected/9663e76d-9ba9-496c-a654-7493c2aa1268-kube-api-access-4d5br\") pod \"olm-operator-6b444d44fb-t4wbk\" (UID: \"9663e76d-9ba9-496c-a654-7493c2aa1268\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467927 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-registry-tls\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467947 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-plugins-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467967 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7aede9e9-6b07-4959-8097-982b17d5628d-profile-collector-cert\") pod \"catalog-operator-68c6474976-6d9nh\" (UID: \"7aede9e9-6b07-4959-8097-982b17d5628d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.467984 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pzcz\" (UniqueName: \"kubernetes.io/projected/7aede9e9-6b07-4959-8097-982b17d5628d-kube-api-access-7pzcz\") pod \"catalog-operator-68c6474976-6d9nh\" (UID: \"7aede9e9-6b07-4959-8097-982b17d5628d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468003 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22faf60d-16a2-49ae-9eb5-5f808013081e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468036 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/17a805ea-4e7f-4204-9712-8be2c673363d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468053 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e116addf-dac0-42aa-9c0e-82383b214399-secret-volume\") pod \"collect-profiles-29399445-2gfhd\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468075 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468094 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fde84f5d-c045-4a70-9d7d-9f4ca71d4351-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-knplv\" (UID: \"fde84f5d-c045-4a70-9d7d-9f4ca71d4351\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468115 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66a7f09b-b285-4635-8666-b8caaa05b1c7-config-volume\") pod \"dns-default-w4d6j\" (UID: \"66a7f09b-b285-4635-8666-b8caaa05b1c7\") " pod="openshift-dns/dns-default-w4d6j" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468134 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468153 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0625b922-f42b-4378-ba25-1ee3fc85cce6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2npgz\" (UID: \"0625b922-f42b-4378-ba25-1ee3fc85cce6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468171 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-tmpfs\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468187 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e51884d-3817-4148-8556-5d91843c2e91-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d9jlc\" (UID: \"9e51884d-3817-4148-8556-5d91843c2e91\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468212 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-trusted-ca\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468233 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6pv6\" (UniqueName: \"kubernetes.io/projected/61518cfd-2e9e-4d2f-95ed-612afa636cef-kube-api-access-r6pv6\") pod \"multus-admission-controller-857f4d67dd-g4h5t\" (UID: \"61518cfd-2e9e-4d2f-95ed-612afa636cef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468268 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/22faf60d-16a2-49ae-9eb5-5f808013081e-audit-dir\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468287 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-apiservice-cert\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468303 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e51884d-3817-4148-8556-5d91843c2e91-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d9jlc\" (UID: \"9e51884d-3817-4148-8556-5d91843c2e91\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468325 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntwk4\" (UniqueName: \"kubernetes.io/projected/9e51884d-3817-4148-8556-5d91843c2e91-kube-api-access-ntwk4\") pod \"kube-storage-version-migrator-operator-b67b599dd-d9jlc\" (UID: \"9e51884d-3817-4148-8556-5d91843c2e91\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468354 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdt74\" (UniqueName: \"kubernetes.io/projected/31d8b755-9373-401c-8663-dba43a056750-kube-api-access-bdt74\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468382 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-dir\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468401 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b755-9373-401c-8663-dba43a056750-auth-proxy-config\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468423 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/626e983a-0392-49ce-9b4b-96c2db7320eb-config\") pod \"service-ca-operator-777779d784-64b4w\" (UID: \"626e983a-0392-49ce-9b4b-96c2db7320eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468447 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndpp6\" (UniqueName: \"kubernetes.io/projected/7edcb436-bf5b-4afd-9603-4a8131b2be16-kube-api-access-ndpp6\") pod \"ingress-canary-27fxx\" (UID: \"7edcb436-bf5b-4afd-9603-4a8131b2be16\") " pod="openshift-ingress-canary/ingress-canary-27fxx" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468486 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtzdz\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-kube-api-access-gtzdz\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468511 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pcvtw\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468531 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-bound-sa-token\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468548 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468565 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0625b922-f42b-4378-ba25-1ee3fc85cce6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2npgz\" (UID: \"0625b922-f42b-4378-ba25-1ee3fc85cce6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468585 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b7761980-8663-4dda-924b-2fb787fcdac8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.468617 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4708bfd8-5610-47ff-a842-d2bc1c6e05bc-proxy-tls\") pod \"machine-config-controller-84d6567774-n2sth\" (UID: \"4708bfd8-5610-47ff-a842-d2bc1c6e05bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.469979 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/22faf60d-16a2-49ae-9eb5-5f808013081e-audit-dir\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.470019 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/892c2271-ba94-41b9-b5ed-04d45c2b1194-proxy-tls\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.470056 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttmm7\" (UniqueName: \"kubernetes.io/projected/5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f-kube-api-access-ttmm7\") pod \"machine-config-server-dv8qs\" (UID: \"5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f\") " pod="openshift-machine-config-operator/machine-config-server-dv8qs" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.470082 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgwr6\" (UniqueName: \"kubernetes.io/projected/892c2271-ba94-41b9-b5ed-04d45c2b1194-kube-api-access-mgwr6\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.470104 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.470127 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6bgh\" (UniqueName: \"kubernetes.io/projected/7dc41818-1f5e-4472-a726-f1322a1f462b-kube-api-access-l6bgh\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.470175 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f394c75-c5d2-4b2a-8845-d2337bc1735e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jf7t\" (UID: \"7f394c75-c5d2-4b2a-8845-d2337bc1735e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.470198 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nf9tt\" (UID: \"c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.470217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dmbv\" (UniqueName: \"kubernetes.io/projected/626e983a-0392-49ce-9b4b-96c2db7320eb-kube-api-access-2dmbv\") pod \"service-ca-operator-777779d784-64b4w\" (UID: \"626e983a-0392-49ce-9b4b-96c2db7320eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.470242 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw9sc\" (UniqueName: \"kubernetes.io/projected/22faf60d-16a2-49ae-9eb5-5f808013081e-kube-api-access-lw9sc\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.470956 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw8t8\" (UniqueName: \"kubernetes.io/projected/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-kube-api-access-cw8t8\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.470979 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-registration-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471012 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f394c75-c5d2-4b2a-8845-d2337bc1735e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jf7t\" (UID: \"7f394c75-c5d2-4b2a-8845-d2337bc1735e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471036 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/22faf60d-16a2-49ae-9eb5-5f808013081e-etcd-client\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471053 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/22faf60d-16a2-49ae-9eb5-5f808013081e-encryption-config\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471074 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0625b922-f42b-4378-ba25-1ee3fc85cce6-config\") pod \"kube-controller-manager-operator-78b949d7b-2npgz\" (UID: \"0625b922-f42b-4378-ba25-1ee3fc85cce6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471090 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-webhook-cert\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471120 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471137 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b9m8\" (UniqueName: \"kubernetes.io/projected/5e8415c7-91a0-4d83-9627-d7e394393334-kube-api-access-6b9m8\") pod \"migrator-59844c95c7-79kln\" (UID: \"5e8415c7-91a0-4d83-9627-d7e394393334\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471154 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9663e76d-9ba9-496c-a654-7493c2aa1268-srv-cert\") pod \"olm-operator-6b444d44fb-t4wbk\" (UID: \"9663e76d-9ba9-496c-a654-7493c2aa1268\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471175 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61518cfd-2e9e-4d2f-95ed-612afa636cef-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-g4h5t\" (UID: \"61518cfd-2e9e-4d2f-95ed-612afa636cef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471194 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9663e76d-9ba9-496c-a654-7493c2aa1268-profile-collector-cert\") pod \"olm-operator-6b444d44fb-t4wbk\" (UID: \"9663e76d-9ba9-496c-a654-7493c2aa1268\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471228 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pcvtw\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471248 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntd5q\" (UniqueName: \"kubernetes.io/projected/7776e83d-dc11-4222-8440-cf059292e127-kube-api-access-ntd5q\") pod \"package-server-manager-789f6589d5-2pchl\" (UID: \"7776e83d-dc11-4222-8440-cf059292e127\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471269 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/626e983a-0392-49ce-9b4b-96c2db7320eb-serving-cert\") pod \"service-ca-operator-777779d784-64b4w\" (UID: \"626e983a-0392-49ce-9b4b-96c2db7320eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471290 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f394c75-c5d2-4b2a-8845-d2337bc1735e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jf7t\" (UID: \"7f394c75-c5d2-4b2a-8845-d2337bc1735e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471322 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471339 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31d8b755-9373-401c-8663-dba43a056750-config\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471359 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckskt\" (UniqueName: \"kubernetes.io/projected/baecc37e-a64f-449f-bde9-f4ff3a66e628-kube-api-access-ckskt\") pod \"service-ca-9c57cc56f-bbm7w\" (UID: \"baecc37e-a64f-449f-bde9-f4ff3a66e628\") " pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471377 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7776e83d-dc11-4222-8440-cf059292e127-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2pchl\" (UID: \"7776e83d-dc11-4222-8440-cf059292e127\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471396 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-socket-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471444 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471464 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/892c2271-ba94-41b9-b5ed-04d45c2b1194-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471485 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-registry-certificates\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471504 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/66a7f09b-b285-4635-8666-b8caaa05b1c7-metrics-tls\") pod \"dns-default-w4d6j\" (UID: \"66a7f09b-b285-4635-8666-b8caaa05b1c7\") " pod="openshift-dns/dns-default-w4d6j" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471537 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/baecc37e-a64f-449f-bde9-f4ff3a66e628-signing-cabundle\") pod \"service-ca-9c57cc56f-bbm7w\" (UID: \"baecc37e-a64f-449f-bde9-f4ff3a66e628\") " pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.471561 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8kb6\" (UniqueName: \"kubernetes.io/projected/4708bfd8-5610-47ff-a842-d2bc1c6e05bc-kube-api-access-g8kb6\") pod \"machine-config-controller-84d6567774-n2sth\" (UID: \"4708bfd8-5610-47ff-a842-d2bc1c6e05bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.472813 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17a805ea-4e7f-4204-9712-8be2c673363d-trusted-ca\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.473007 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-dir\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.473406 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b755-9373-401c-8663-dba43a056750-auth-proxy-config\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.474794 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.475362 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/22faf60d-16a2-49ae-9eb5-5f808013081e-audit-policies\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.476097 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.476502 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-policies\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.476624 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22faf60d-16a2-49ae-9eb5-5f808013081e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.476777 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.477394 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4708bfd8-5610-47ff-a842-d2bc1c6e05bc-proxy-tls\") pod \"machine-config-controller-84d6567774-n2sth\" (UID: \"4708bfd8-5610-47ff-a842-d2bc1c6e05bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" Nov 24 06:49:57 crc kubenswrapper[4799]: E1124 06:49:57.477419 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:57.977391146 +0000 UTC m=+143.633373620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.477571 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/892c2271-ba94-41b9-b5ed-04d45c2b1194-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.478204 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-registry-tls\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.479113 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b7761980-8663-4dda-924b-2fb787fcdac8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.480142 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-registry-certificates\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.482602 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0625b922-f42b-4378-ba25-1ee3fc85cce6-config\") pod \"kube-controller-manager-operator-78b949d7b-2npgz\" (UID: \"0625b922-f42b-4378-ba25-1ee3fc85cce6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.483207 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.483318 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31d8b755-9373-401c-8663-dba43a056750-config\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.484567 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-trusted-ca\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.485218 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/22faf60d-16a2-49ae-9eb5-5f808013081e-encryption-config\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.485609 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.486094 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/31d8b755-9373-401c-8663-dba43a056750-machine-approver-tls\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.486787 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.486979 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.490697 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fde84f5d-c045-4a70-9d7d-9f4ca71d4351-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-knplv\" (UID: \"fde84f5d-c045-4a70-9d7d-9f4ca71d4351\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.493237 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f394c75-c5d2-4b2a-8845-d2337bc1735e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jf7t\" (UID: \"7f394c75-c5d2-4b2a-8845-d2337bc1735e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.493379 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nf9tt\" (UID: \"c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.494131 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/22faf60d-16a2-49ae-9eb5-5f808013081e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.506083 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f394c75-c5d2-4b2a-8845-d2337bc1735e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jf7t\" (UID: \"7f394c75-c5d2-4b2a-8845-d2337bc1735e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.507129 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.507156 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61518cfd-2e9e-4d2f-95ed-612afa636cef-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-g4h5t\" (UID: \"61518cfd-2e9e-4d2f-95ed-612afa636cef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.507253 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.508284 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/892c2271-ba94-41b9-b5ed-04d45c2b1194-proxy-tls\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.521617 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47dvt\" (UniqueName: \"kubernetes.io/projected/17a805ea-4e7f-4204-9712-8be2c673363d-kube-api-access-47dvt\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.521802 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0625b922-f42b-4378-ba25-1ee3fc85cce6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2npgz\" (UID: \"0625b922-f42b-4378-ba25-1ee3fc85cce6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.521935 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/22faf60d-16a2-49ae-9eb5-5f808013081e-etcd-client\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.526839 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.562254 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" event={"ID":"0119fbf1-7180-4670-a389-adcf32d1a736","Type":"ContainerStarted","Data":"33529e41966f0bb6b8050dc8b5c53499e017bd77197dfe6074f648b462521fae"} Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.563053 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmbs6\" (UniqueName: \"kubernetes.io/projected/d6a12d20-1525-4cf5-8e5c-da7184c42581-kube-api-access-zmbs6\") pod \"downloads-7954f5f757-6j6bn\" (UID: \"d6a12d20-1525-4cf5-8e5c-da7184c42581\") " pod="openshift-console/downloads-7954f5f757-6j6bn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.564451 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" event={"ID":"12dc0898-d8d0-4b23-a8e2-3ed45987583b","Type":"ContainerStarted","Data":"c7debf7cf98f2a33908ae7b442db27e16375685a4de41d44e29d7cafcb0ffd6c"} Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.566732 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cn7ch" event={"ID":"1317991a-3615-4dc5-9cbe-232e04180a7c","Type":"ContainerStarted","Data":"1c567ee0bf509302ff7b2621b8b750e1bbe8f593acc7c79fcd1459dcf5771353"} Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.567737 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zzd4b" event={"ID":"a63db911-38e5-47a8-926a-5181ffa957bc","Type":"ContainerStarted","Data":"c51b5c7645f10bfa913153da514ab85abd8c3475594e0bb2902b9d416cd5ef58"} Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.569486 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8kb6\" (UniqueName: \"kubernetes.io/projected/4708bfd8-5610-47ff-a842-d2bc1c6e05bc-kube-api-access-g8kb6\") pod \"machine-config-controller-84d6567774-n2sth\" (UID: \"4708bfd8-5610-47ff-a842-d2bc1c6e05bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573321 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573695 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/66a7f09b-b285-4635-8666-b8caaa05b1c7-metrics-tls\") pod \"dns-default-w4d6j\" (UID: \"66a7f09b-b285-4635-8666-b8caaa05b1c7\") " pod="openshift-dns/dns-default-w4d6j" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573733 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/baecc37e-a64f-449f-bde9-f4ff3a66e628-signing-cabundle\") pod \"service-ca-9c57cc56f-bbm7w\" (UID: \"baecc37e-a64f-449f-bde9-f4ff3a66e628\") " pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573762 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7edcb436-bf5b-4afd-9603-4a8131b2be16-cert\") pod \"ingress-canary-27fxx\" (UID: \"7edcb436-bf5b-4afd-9603-4a8131b2be16\") " pod="openshift-ingress-canary/ingress-canary-27fxx" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573787 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwdqk\" (UniqueName: \"kubernetes.io/projected/e116addf-dac0-42aa-9c0e-82383b214399-kube-api-access-mwdqk\") pod \"collect-profiles-29399445-2gfhd\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573818 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzrrh\" (UniqueName: \"kubernetes.io/projected/3b02c876-c14b-4422-beaf-9ea3ab910f90-kube-api-access-xzrrh\") pod \"marketplace-operator-79b997595-pcvtw\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573856 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/baecc37e-a64f-449f-bde9-f4ff3a66e628-signing-key\") pod \"service-ca-9c57cc56f-bbm7w\" (UID: \"baecc37e-a64f-449f-bde9-f4ff3a66e628\") " pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573888 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-mountpoint-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573911 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-csi-data-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573935 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f-node-bootstrap-token\") pod \"machine-config-server-dv8qs\" (UID: \"5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f\") " pod="openshift-machine-config-operator/machine-config-server-dv8qs" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573960 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f-certs\") pod \"machine-config-server-dv8qs\" (UID: \"5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f\") " pod="openshift-machine-config-operator/machine-config-server-dv8qs" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.573992 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57pv7\" (UniqueName: \"kubernetes.io/projected/2927bbfb-073e-41b2-bb65-c6c75a54560b-kube-api-access-57pv7\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574016 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7aede9e9-6b07-4959-8097-982b17d5628d-srv-cert\") pod \"catalog-operator-68c6474976-6d9nh\" (UID: \"7aede9e9-6b07-4959-8097-982b17d5628d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574065 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g8z8\" (UniqueName: \"kubernetes.io/projected/66a7f09b-b285-4635-8666-b8caaa05b1c7-kube-api-access-2g8z8\") pod \"dns-default-w4d6j\" (UID: \"66a7f09b-b285-4635-8666-b8caaa05b1c7\") " pod="openshift-dns/dns-default-w4d6j" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574092 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e116addf-dac0-42aa-9c0e-82383b214399-config-volume\") pod \"collect-profiles-29399445-2gfhd\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574141 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d5br\" (UniqueName: \"kubernetes.io/projected/9663e76d-9ba9-496c-a654-7493c2aa1268-kube-api-access-4d5br\") pod \"olm-operator-6b444d44fb-t4wbk\" (UID: \"9663e76d-9ba9-496c-a654-7493c2aa1268\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574168 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-plugins-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574195 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7aede9e9-6b07-4959-8097-982b17d5628d-profile-collector-cert\") pod \"catalog-operator-68c6474976-6d9nh\" (UID: \"7aede9e9-6b07-4959-8097-982b17d5628d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pzcz\" (UniqueName: \"kubernetes.io/projected/7aede9e9-6b07-4959-8097-982b17d5628d-kube-api-access-7pzcz\") pod \"catalog-operator-68c6474976-6d9nh\" (UID: \"7aede9e9-6b07-4959-8097-982b17d5628d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:49:57 crc kubenswrapper[4799]: E1124 06:49:57.574255 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:58.074230647 +0000 UTC m=+143.730213131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574288 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e116addf-dac0-42aa-9c0e-82383b214399-secret-volume\") pod \"collect-profiles-29399445-2gfhd\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574317 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66a7f09b-b285-4635-8666-b8caaa05b1c7-config-volume\") pod \"dns-default-w4d6j\" (UID: \"66a7f09b-b285-4635-8666-b8caaa05b1c7\") " pod="openshift-dns/dns-default-w4d6j" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574351 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-tmpfs\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574378 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e51884d-3817-4148-8556-5d91843c2e91-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d9jlc\" (UID: \"9e51884d-3817-4148-8556-5d91843c2e91\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-apiservice-cert\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574440 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e51884d-3817-4148-8556-5d91843c2e91-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d9jlc\" (UID: \"9e51884d-3817-4148-8556-5d91843c2e91\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574464 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntwk4\" (UniqueName: \"kubernetes.io/projected/9e51884d-3817-4148-8556-5d91843c2e91-kube-api-access-ntwk4\") pod \"kube-storage-version-migrator-operator-b67b599dd-d9jlc\" (UID: \"9e51884d-3817-4148-8556-5d91843c2e91\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574510 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/626e983a-0392-49ce-9b4b-96c2db7320eb-config\") pod \"service-ca-operator-777779d784-64b4w\" (UID: \"626e983a-0392-49ce-9b4b-96c2db7320eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574537 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndpp6\" (UniqueName: \"kubernetes.io/projected/7edcb436-bf5b-4afd-9603-4a8131b2be16-kube-api-access-ndpp6\") pod \"ingress-canary-27fxx\" (UID: \"7edcb436-bf5b-4afd-9603-4a8131b2be16\") " pod="openshift-ingress-canary/ingress-canary-27fxx" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574582 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pcvtw\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574626 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttmm7\" (UniqueName: \"kubernetes.io/projected/5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f-kube-api-access-ttmm7\") pod \"machine-config-server-dv8qs\" (UID: \"5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f\") " pod="openshift-machine-config-operator/machine-config-server-dv8qs" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574689 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dmbv\" (UniqueName: \"kubernetes.io/projected/626e983a-0392-49ce-9b4b-96c2db7320eb-kube-api-access-2dmbv\") pod \"service-ca-operator-777779d784-64b4w\" (UID: \"626e983a-0392-49ce-9b4b-96c2db7320eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574727 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-registration-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574752 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw8t8\" (UniqueName: \"kubernetes.io/projected/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-kube-api-access-cw8t8\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574776 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-webhook-cert\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574810 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9663e76d-9ba9-496c-a654-7493c2aa1268-srv-cert\") pod \"olm-operator-6b444d44fb-t4wbk\" (UID: \"9663e76d-9ba9-496c-a654-7493c2aa1268\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574837 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9663e76d-9ba9-496c-a654-7493c2aa1268-profile-collector-cert\") pod \"olm-operator-6b444d44fb-t4wbk\" (UID: \"9663e76d-9ba9-496c-a654-7493c2aa1268\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574886 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pcvtw\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574913 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntd5q\" (UniqueName: \"kubernetes.io/projected/7776e83d-dc11-4222-8440-cf059292e127-kube-api-access-ntd5q\") pod \"package-server-manager-789f6589d5-2pchl\" (UID: \"7776e83d-dc11-4222-8440-cf059292e127\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574939 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/626e983a-0392-49ce-9b4b-96c2db7320eb-serving-cert\") pod \"service-ca-operator-777779d784-64b4w\" (UID: \"626e983a-0392-49ce-9b4b-96c2db7320eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574969 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckskt\" (UniqueName: \"kubernetes.io/projected/baecc37e-a64f-449f-bde9-f4ff3a66e628-kube-api-access-ckskt\") pod \"service-ca-9c57cc56f-bbm7w\" (UID: \"baecc37e-a64f-449f-bde9-f4ff3a66e628\") " pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574993 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7776e83d-dc11-4222-8440-cf059292e127-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2pchl\" (UID: \"7776e83d-dc11-4222-8440-cf059292e127\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.575015 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-socket-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.575380 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-socket-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.576878 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e116addf-dac0-42aa-9c0e-82383b214399-config-volume\") pod \"collect-profiles-29399445-2gfhd\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574090 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-mountpoint-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.577223 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-plugins-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.578175 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e51884d-3817-4148-8556-5d91843c2e91-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d9jlc\" (UID: \"9e51884d-3817-4148-8556-5d91843c2e91\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.578997 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/baecc37e-a64f-449f-bde9-f4ff3a66e628-signing-cabundle\") pod \"service-ca-9c57cc56f-bbm7w\" (UID: \"baecc37e-a64f-449f-bde9-f4ff3a66e628\") " pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.579368 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/66a7f09b-b285-4635-8666-b8caaa05b1c7-metrics-tls\") pod \"dns-default-w4d6j\" (UID: \"66a7f09b-b285-4635-8666-b8caaa05b1c7\") " pod="openshift-dns/dns-default-w4d6j" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.574417 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-csi-data-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.581395 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2927bbfb-073e-41b2-bb65-c6c75a54560b-registration-dir\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.581805 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e51884d-3817-4148-8556-5d91843c2e91-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d9jlc\" (UID: \"9e51884d-3817-4148-8556-5d91843c2e91\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.581999 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-tmpfs\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.582028 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/626e983a-0392-49ce-9b4b-96c2db7320eb-config\") pod \"service-ca-operator-777779d784-64b4w\" (UID: \"626e983a-0392-49ce-9b4b-96c2db7320eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.582253 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66a7f09b-b285-4635-8666-b8caaa05b1c7-config-volume\") pod \"dns-default-w4d6j\" (UID: \"66a7f09b-b285-4635-8666-b8caaa05b1c7\") " pod="openshift-dns/dns-default-w4d6j" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.582651 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/baecc37e-a64f-449f-bde9-f4ff3a66e628-signing-key\") pod \"service-ca-9c57cc56f-bbm7w\" (UID: \"baecc37e-a64f-449f-bde9-f4ff3a66e628\") " pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.583096 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pcvtw\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.583369 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-apiservice-cert\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.583524 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9663e76d-9ba9-496c-a654-7493c2aa1268-srv-cert\") pod \"olm-operator-6b444d44fb-t4wbk\" (UID: \"9663e76d-9ba9-496c-a654-7493c2aa1268\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.584047 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7aede9e9-6b07-4959-8097-982b17d5628d-srv-cert\") pod \"catalog-operator-68c6474976-6d9nh\" (UID: \"7aede9e9-6b07-4959-8097-982b17d5628d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.584923 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdt74\" (UniqueName: \"kubernetes.io/projected/31d8b755-9373-401c-8663-dba43a056750-kube-api-access-bdt74\") pod \"machine-approver-56656f9798-dwshf\" (UID: \"31d8b755-9373-401c-8663-dba43a056750\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.585061 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f-node-bootstrap-token\") pod \"machine-config-server-dv8qs\" (UID: \"5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f\") " pod="openshift-machine-config-operator/machine-config-server-dv8qs" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.585566 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pcvtw\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.586699 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f-certs\") pod \"machine-config-server-dv8qs\" (UID: \"5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f\") " pod="openshift-machine-config-operator/machine-config-server-dv8qs" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.587084 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-webhook-cert\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.587725 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/626e983a-0392-49ce-9b4b-96c2db7320eb-serving-cert\") pod \"service-ca-operator-777779d784-64b4w\" (UID: \"626e983a-0392-49ce-9b4b-96c2db7320eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.587969 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9663e76d-9ba9-496c-a654-7493c2aa1268-profile-collector-cert\") pod \"olm-operator-6b444d44fb-t4wbk\" (UID: \"9663e76d-9ba9-496c-a654-7493c2aa1268\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.588407 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7aede9e9-6b07-4959-8097-982b17d5628d-profile-collector-cert\") pod \"catalog-operator-68c6474976-6d9nh\" (UID: \"7aede9e9-6b07-4959-8097-982b17d5628d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.589250 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7776e83d-dc11-4222-8440-cf059292e127-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2pchl\" (UID: \"7776e83d-dc11-4222-8440-cf059292e127\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.590484 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e116addf-dac0-42aa-9c0e-82383b214399-secret-volume\") pod \"collect-profiles-29399445-2gfhd\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.590881 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7edcb436-bf5b-4afd-9603-4a8131b2be16-cert\") pod \"ingress-canary-27fxx\" (UID: \"7edcb436-bf5b-4afd-9603-4a8131b2be16\") " pod="openshift-ingress-canary/ingress-canary-27fxx" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.593690 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6j6bn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.604205 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.607466 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtzdz\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-kube-api-access-gtzdz\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.630424 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkpjn\" (UniqueName: \"kubernetes.io/projected/fde84f5d-c045-4a70-9d7d-9f4ca71d4351-kube-api-access-gkpjn\") pod \"control-plane-machine-set-operator-78cbb6b69f-knplv\" (UID: \"fde84f5d-c045-4a70-9d7d-9f4ca71d4351\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.654034 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-bound-sa-token\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.669753 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/17a805ea-4e7f-4204-9712-8be2c673363d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pxzr4\" (UID: \"17a805ea-4e7f-4204-9712-8be2c673363d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.676622 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: E1124 06:49:57.677189 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:58.177148413 +0000 UTC m=+143.833130887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.691907 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgwr6\" (UniqueName: \"kubernetes.io/projected/892c2271-ba94-41b9-b5ed-04d45c2b1194-kube-api-access-mgwr6\") pod \"machine-config-operator-74547568cd-9wp9k\" (UID: \"892c2271-ba94-41b9-b5ed-04d45c2b1194\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.698903 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.706760 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nf9tt\" (UID: \"c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.732788 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rzh22"] Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.732884 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.747190 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6bgh\" (UniqueName: \"kubernetes.io/projected/7dc41818-1f5e-4472-a726-f1322a1f462b-kube-api-access-l6bgh\") pod \"oauth-openshift-558db77b4-7gmz9\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.748891 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.749450 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nqlvh"] Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.756057 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.759658 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f394c75-c5d2-4b2a-8845-d2337bc1735e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jf7t\" (UID: \"7f394c75-c5d2-4b2a-8845-d2337bc1735e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.763141 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw9sc\" (UniqueName: \"kubernetes.io/projected/22faf60d-16a2-49ae-9eb5-5f808013081e-kube-api-access-lw9sc\") pod \"apiserver-7bbb656c7d-wsxhz\" (UID: \"22faf60d-16a2-49ae-9eb5-5f808013081e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.765193 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qv7f9"] Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.778351 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:57 crc kubenswrapper[4799]: E1124 06:49:57.780486 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:58.280458172 +0000 UTC m=+143.936440646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.781665 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: E1124 06:49:57.782126 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:58.282114518 +0000 UTC m=+143.938096992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.787414 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b9m8\" (UniqueName: \"kubernetes.io/projected/5e8415c7-91a0-4d83-9627-d7e394393334-kube-api-access-6b9m8\") pod \"migrator-59844c95c7-79kln\" (UID: \"5e8415c7-91a0-4d83-9627-d7e394393334\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.810813 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0625b922-f42b-4378-ba25-1ee3fc85cce6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2npgz\" (UID: \"0625b922-f42b-4378-ba25-1ee3fc85cce6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.827284 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6pv6\" (UniqueName: \"kubernetes.io/projected/61518cfd-2e9e-4d2f-95ed-612afa636cef-kube-api-access-r6pv6\") pod \"multus-admission-controller-857f4d67dd-g4h5t\" (UID: \"61518cfd-2e9e-4d2f-95ed-612afa636cef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.873748 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwdqk\" (UniqueName: \"kubernetes.io/projected/e116addf-dac0-42aa-9c0e-82383b214399-kube-api-access-mwdqk\") pod \"collect-profiles-29399445-2gfhd\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.882463 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:57 crc kubenswrapper[4799]: E1124 06:49:57.883231 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:58.383202903 +0000 UTC m=+144.039185377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.887576 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzrrh\" (UniqueName: \"kubernetes.io/projected/3b02c876-c14b-4422-beaf-9ea3ab910f90-kube-api-access-xzrrh\") pod \"marketplace-operator-79b997595-pcvtw\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.914619 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pzcz\" (UniqueName: \"kubernetes.io/projected/7aede9e9-6b07-4959-8097-982b17d5628d-kube-api-access-7pzcz\") pod \"catalog-operator-68c6474976-6d9nh\" (UID: \"7aede9e9-6b07-4959-8097-982b17d5628d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.930241 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57pv7\" (UniqueName: \"kubernetes.io/projected/2927bbfb-073e-41b2-bb65-c6c75a54560b-kube-api-access-57pv7\") pod \"csi-hostpathplugin-2zg6b\" (UID: \"2927bbfb-073e-41b2-bb65-c6c75a54560b\") " pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.940262 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.944293 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g8z8\" (UniqueName: \"kubernetes.io/projected/66a7f09b-b285-4635-8666-b8caaa05b1c7-kube-api-access-2g8z8\") pod \"dns-default-w4d6j\" (UID: \"66a7f09b-b285-4635-8666-b8caaa05b1c7\") " pod="openshift-dns/dns-default-w4d6j" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.962477 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d5br\" (UniqueName: \"kubernetes.io/projected/9663e76d-9ba9-496c-a654-7493c2aa1268-kube-api-access-4d5br\") pod \"olm-operator-6b444d44fb-t4wbk\" (UID: \"9663e76d-9ba9-496c-a654-7493c2aa1268\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.971947 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.972267 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.987885 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.988379 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.992536 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:57 crc kubenswrapper[4799]: I1124 06:49:57.995500 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndpp6\" (UniqueName: \"kubernetes.io/projected/7edcb436-bf5b-4afd-9603-4a8131b2be16-kube-api-access-ndpp6\") pod \"ingress-canary-27fxx\" (UID: \"7edcb436-bf5b-4afd-9603-4a8131b2be16\") " pod="openshift-ingress-canary/ingress-canary-27fxx" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.009325 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.014701 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntwk4\" (UniqueName: \"kubernetes.io/projected/9e51884d-3817-4148-8556-5d91843c2e91-kube-api-access-ntwk4\") pod \"kube-storage-version-migrator-operator-b67b599dd-d9jlc\" (UID: \"9e51884d-3817-4148-8556-5d91843c2e91\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" Nov 24 06:49:58 crc kubenswrapper[4799]: E1124 06:49:58.022054 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:58.522029382 +0000 UTC m=+144.178011856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.040898 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.042230 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-b28hc"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.042971 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.058161 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttmm7\" (UniqueName: \"kubernetes.io/projected/5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f-kube-api-access-ttmm7\") pod \"machine-config-server-dv8qs\" (UID: \"5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f\") " pod="openshift-machine-config-operator/machine-config-server-dv8qs" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.061231 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckskt\" (UniqueName: \"kubernetes.io/projected/baecc37e-a64f-449f-bde9-f4ff3a66e628-kube-api-access-ckskt\") pod \"service-ca-9c57cc56f-bbm7w\" (UID: \"baecc37e-a64f-449f-bde9-f4ff3a66e628\") " pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.067650 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.077599 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.082159 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dmbv\" (UniqueName: \"kubernetes.io/projected/626e983a-0392-49ce-9b4b-96c2db7320eb-kube-api-access-2dmbv\") pod \"service-ca-operator-777779d784-64b4w\" (UID: \"626e983a-0392-49ce-9b4b-96c2db7320eb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.085274 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.087657 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw8t8\" (UniqueName: \"kubernetes.io/projected/a3fa15d5-b246-4af2-9fbc-55bc4d5de8af-kube-api-access-cw8t8\") pod \"packageserver-d55dfcdfc-qn45g\" (UID: \"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.093574 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:58 crc kubenswrapper[4799]: E1124 06:49:58.094161 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:58.594138327 +0000 UTC m=+144.250120801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.094331 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.102154 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntd5q\" (UniqueName: \"kubernetes.io/projected/7776e83d-dc11-4222-8440-cf059292e127-kube-api-access-ntd5q\") pod \"package-server-manager-789f6589d5-2pchl\" (UID: \"7776e83d-dc11-4222-8440-cf059292e127\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.106239 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.113318 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.121430 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.129761 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.138239 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.146689 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.171593 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.175308 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-27fxx" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.187362 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-dv8qs" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.194962 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:58 crc kubenswrapper[4799]: E1124 06:49:58.195401 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:58.695383777 +0000 UTC m=+144.351366251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.200951 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-w4d6j" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.224919 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-glxff"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.244523 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.245883 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rhhsg"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.275175 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.289446 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6j6bn"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.296417 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:58 crc kubenswrapper[4799]: E1124 06:49:58.296583 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:58.796560425 +0000 UTC m=+144.452542899 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.297052 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:58 crc kubenswrapper[4799]: E1124 06:49:58.297649 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:58.79761842 +0000 UTC m=+144.453600894 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.340724 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t6hw2"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.351429 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-58mkm"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.397982 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:58 crc kubenswrapper[4799]: E1124 06:49:58.398584 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:58.89856084 +0000 UTC m=+144.554543314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.457908 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.475008 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth"] Nov 24 06:49:58 crc kubenswrapper[4799]: W1124 06:49:58.479148 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod336a419e_e807_4859_b624_2ed06a9a8665.slice/crio-909b61f4a543f5f2fe4561491241034a8d45e42e797990ba6522d93f744f88ea WatchSource:0}: Error finding container 909b61f4a543f5f2fe4561491241034a8d45e42e797990ba6522d93f744f88ea: Status 404 returned error can't find the container with id 909b61f4a543f5f2fe4561491241034a8d45e42e797990ba6522d93f744f88ea Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.501553 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:58 crc kubenswrapper[4799]: E1124 06:49:58.502035 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:59.002021944 +0000 UTC m=+144.658004418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.521536 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.538778 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.602784 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:58 crc kubenswrapper[4799]: E1124 06:49:58.603412 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:59.103387718 +0000 UTC m=+144.759370192 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.608644 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gmz9"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.617986 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.641239 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zzd4b" event={"ID":"a63db911-38e5-47a8-926a-5181ffa957bc","Type":"ContainerStarted","Data":"95ba7136a5c4b0bbf12987d35736c8273b35d29af888440a1511483b9c90869e"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.642071 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.644639 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" event={"ID":"31d8b755-9373-401c-8663-dba43a056750","Type":"ContainerStarted","Data":"ca7f6da4a00acf4300a5dd0ad710c2ae3636b2a559e3e7f1efbf5289a5a64a3b"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.644674 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" event={"ID":"31d8b755-9373-401c-8663-dba43a056750","Type":"ContainerStarted","Data":"5b5fc5bc6af8c98d7a98813d92ce09c489261e30371f7090c89b6062962ad7ee"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.646811 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" event={"ID":"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f","Type":"ContainerStarted","Data":"0dd5a412f7e81a8162c718119217d92ce7898c318b55dc3358bb26d1411b89de"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.648099 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" event={"ID":"ed31a762-47f8-4efa-88f4-262e93cbdcea","Type":"ContainerStarted","Data":"9a5936234f3a0eb4126c980d9df5daf2ad26cf123acf1cefa10109c0a5cc1ea2"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.649062 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.655428 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cn7ch" event={"ID":"1317991a-3615-4dc5-9cbe-232e04180a7c","Type":"ContainerStarted","Data":"e670b3459fb8e4173cf353be649d1846b63bc4783d3e01a0abbf6407f60fc83a"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.660693 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.675973 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.682266 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" event={"ID":"12dc0898-d8d0-4b23-a8e2-3ed45987583b","Type":"ContainerStarted","Data":"6a932f355fdc8674f7a0fd65f0b4d0002120a955680224056bf5e8bf6d83f731"} Nov 24 06:49:58 crc kubenswrapper[4799]: W1124 06:49:58.691198 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfde84f5d_c045_4a70_9d7d_9f4ca71d4351.slice/crio-5df1347c0c3755ecc42eb4d10d291f363f5b1eea77417578b27a0f2e774cdaff WatchSource:0}: Error finding container 5df1347c0c3755ecc42eb4d10d291f363f5b1eea77417578b27a0f2e774cdaff: Status 404 returned error can't find the container with id 5df1347c0c3755ecc42eb4d10d291f363f5b1eea77417578b27a0f2e774cdaff Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.696732 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" event={"ID":"90dfb996-eeb4-4004-8879-86be3f6a9e66","Type":"ContainerStarted","Data":"257d98a6e70f5bdb3f2d4ede252aee6653e490fe2a39c054428052ad6a25a49f"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.699287 4799 patch_prober.go:28] interesting pod/console-operator-58897d9998-zzd4b container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.699367 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-zzd4b" podUID="a63db911-38e5-47a8-926a-5181ffa957bc" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.703698 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" event={"ID":"facf7448-3c40-476b-bc7f-9526c5f4c3e2","Type":"ContainerStarted","Data":"a642fa0b64b17ac7827420f59ca84cdcd610f60612c97e98da8d1347f7b5dc62"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.704655 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:58 crc kubenswrapper[4799]: E1124 06:49:58.705723 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:59.205708364 +0000 UTC m=+144.861690838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.712181 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6j6bn" event={"ID":"d6a12d20-1525-4cf5-8e5c-da7184c42581","Type":"ContainerStarted","Data":"e0bfb023215407915b4e317fc1791709d27be2060714082cae93592c9e54e935"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.718494 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" event={"ID":"bfae07f4-0529-426f-aaac-d82964808045","Type":"ContainerStarted","Data":"6f16494cfbb03d45fa698b6b57a51a1911f5a7c984eda9929dbccf7ef702a5de"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.732911 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" event={"ID":"50f64a8f-2611-4be4-b8e4-08eba7fa0e69","Type":"ContainerStarted","Data":"decdc7b52dfef2b1946559bfde91895f304e3549dbc04f8c6bfd4aca4ef3f8a2"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.739767 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" event={"ID":"fec9f210-0d28-463f-b3bf-4f27c6eee903","Type":"ContainerStarted","Data":"918117ab1d5ea1c4d9005653c93ba9e88f0c29e8255fc0ad6aae6417bc33f071"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.760606 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" event={"ID":"36f481dd-d3ab-422a-8f26-69fcccb51410","Type":"ContainerStarted","Data":"179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.760669 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" event={"ID":"36f481dd-d3ab-422a-8f26-69fcccb51410","Type":"ContainerStarted","Data":"0034ea623d86f7e724d60cf08ff7e83b6031880f026455483c99c21fd0951459"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.762101 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.765888 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-58mkm" event={"ID":"336a419e-e807-4859-b624-2ed06a9a8665","Type":"ContainerStarted","Data":"909b61f4a543f5f2fe4561491241034a8d45e42e797990ba6522d93f744f88ea"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.771930 4799 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rzh22 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.771992 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" podUID="36f481dd-d3ab-422a-8f26-69fcccb51410" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.775359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" event={"ID":"86f640cf-7c54-4b44-a52f-21378c535d5a","Type":"ContainerStarted","Data":"d4b918f88876e5483d3d50446ac7be8dd97333e9a7082dda41d290b35649ac89"} Nov 24 06:49:58 crc kubenswrapper[4799]: W1124 06:49:58.790331 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f394c75_c5d2_4b2a_8845_d2337bc1735e.slice/crio-8946123ab04c96eeedb0eda8f4096b812fbf8e4a772ceee2d7c1264b0ebd1790 WatchSource:0}: Error finding container 8946123ab04c96eeedb0eda8f4096b812fbf8e4a772ceee2d7c1264b0ebd1790: Status 404 returned error can't find the container with id 8946123ab04c96eeedb0eda8f4096b812fbf8e4a772ceee2d7c1264b0ebd1790 Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.790347 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" event={"ID":"0119fbf1-7180-4670-a389-adcf32d1a736","Type":"ContainerStarted","Data":"70f08164b92ac29454e823a4261761fff6bbac8948075d58695da9b6a37a143e"} Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.801958 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-g4h5t"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.807899 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.810698 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:58 crc kubenswrapper[4799]: E1124 06:49:58.811931 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:59.311904851 +0000 UTC m=+144.967887325 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.891478 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.913536 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh"] Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.913982 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:58 crc kubenswrapper[4799]: E1124 06:49:58.914906 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:59.414862789 +0000 UTC m=+145.070845263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:58 crc kubenswrapper[4799]: I1124 06:49:58.920268 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-cn7ch" podStartSLOduration=121.920246641 podStartE2EDuration="2m1.920246641s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:58.919464944 +0000 UTC m=+144.575447418" watchObservedRunningTime="2025-11-24 06:49:58.920246641 +0000 UTC m=+144.576229105" Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.015499 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:59 crc kubenswrapper[4799]: E1124 06:49:59.016367 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:59.516344877 +0000 UTC m=+145.172327351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.101904 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcvtw"] Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.164272 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:59 crc kubenswrapper[4799]: E1124 06:49:59.165571 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:59.665536666 +0000 UTC m=+145.321519140 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.266376 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:59 crc kubenswrapper[4799]: E1124 06:49:59.266934 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:59.76691452 +0000 UTC m=+145.422896994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.378666 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:59 crc kubenswrapper[4799]: E1124 06:49:59.380416 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:49:59.880398823 +0000 UTC m=+145.536381297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.386855 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl"] Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.389556 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk"] Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.403257 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.407195 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-zzd4b" podStartSLOduration=123.407175598 podStartE2EDuration="2m3.407175598s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:59.405987617 +0000 UTC m=+145.061970091" watchObservedRunningTime="2025-11-24 06:49:59.407175598 +0000 UTC m=+145.063158072" Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.412396 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:49:59 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:49:59 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:49:59 crc kubenswrapper[4799]: healthz check failed Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.412445 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.469229 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-64b4w"] Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.480728 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:59 crc kubenswrapper[4799]: E1124 06:49:59.481361 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:49:59.981334052 +0000 UTC m=+145.637316526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.529125 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2zg6b"] Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.530910 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-27fxx"] Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.568920 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g"] Nov 24 06:49:59 crc kubenswrapper[4799]: W1124 06:49:59.574658 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7776e83d_dc11_4222_8440_cf059292e127.slice/crio-e40632a459eab330348cabffdd7e6a5e380a8b642a47490aa24353ffe57fd3b8 WatchSource:0}: Error finding container e40632a459eab330348cabffdd7e6a5e380a8b642a47490aa24353ffe57fd3b8: Status 404 returned error can't find the container with id e40632a459eab330348cabffdd7e6a5e380a8b642a47490aa24353ffe57fd3b8 Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.579214 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc"] Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.582429 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:59 crc kubenswrapper[4799]: E1124 06:49:59.582926 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:00.082871282 +0000 UTC m=+145.738853756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.596791 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-w4d6j"] Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.610088 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-bbm7w"] Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.683449 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:59 crc kubenswrapper[4799]: E1124 06:49:59.683723 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:00.183682057 +0000 UTC m=+145.839664521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.684085 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:59 crc kubenswrapper[4799]: E1124 06:49:59.684561 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:00.184545026 +0000 UTC m=+145.840527490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:59 crc kubenswrapper[4799]: W1124 06:49:59.773296 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2927bbfb_073e_41b2_bb65_c6c75a54560b.slice/crio-efe583c357ff6f069bffc6981f9e9c20bda5d6356c3fc2614b2907b12d8726d2 WatchSource:0}: Error finding container efe583c357ff6f069bffc6981f9e9c20bda5d6356c3fc2614b2907b12d8726d2: Status 404 returned error can't find the container with id efe583c357ff6f069bffc6981f9e9c20bda5d6356c3fc2614b2907b12d8726d2 Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.784827 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:49:59 crc kubenswrapper[4799]: E1124 06:49:59.785311 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:00.285285839 +0000 UTC m=+145.941268313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.823659 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" event={"ID":"90dfb996-eeb4-4004-8879-86be3f6a9e66","Type":"ContainerStarted","Data":"547fc8f67d70a7979fbae28872c3a5e301011a8d3eadfee953155fba5bef2e27"} Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.862725 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" event={"ID":"31d8b755-9373-401c-8663-dba43a056750","Type":"ContainerStarted","Data":"57f711361eddaba82041ad4f42d63139288e8a4dccedd67f9a9f8f398bbcb929"} Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.902289 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:49:59 crc kubenswrapper[4799]: E1124 06:49:59.902973 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:00.402940742 +0000 UTC m=+146.058923216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.924279 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" event={"ID":"9b717be8-b446-40c0-a5d9-1de2525168fd","Type":"ContainerStarted","Data":"def181a1d89f91757c0697bfe32fbb0e0e8cfc2ad2c4786aeb1ea907339f1ff8"} Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.924335 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" event={"ID":"9b717be8-b446-40c0-a5d9-1de2525168fd","Type":"ContainerStarted","Data":"290428e65163bf07c3c61a5bd4b7574f091286525b272519620db1852e717307"} Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.927211 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" event={"ID":"9663e76d-9ba9-496c-a654-7493c2aa1268","Type":"ContainerStarted","Data":"77f47ce5cde6abbb041fcc462551c2225601bce3a746118452b730a7854da8c9"} Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.954930 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" event={"ID":"e116addf-dac0-42aa-9c0e-82383b214399","Type":"ContainerStarted","Data":"462aeb8d0ee119d1da5c9d2e27b1fd83726d314733294f99511ab1672114f0ab"} Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.981389 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" event={"ID":"50f64a8f-2611-4be4-b8e4-08eba7fa0e69","Type":"ContainerStarted","Data":"5e85d35be8e8169e00a8470e5eda8ce135914c54b5ce99560d53bae5d712e304"} Nov 24 06:49:59 crc kubenswrapper[4799]: I1124 06:49:59.988144 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" podStartSLOduration=122.988115289 podStartE2EDuration="2m2.988115289s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:49:59.982147247 +0000 UTC m=+145.638129721" watchObservedRunningTime="2025-11-24 06:49:59.988115289 +0000 UTC m=+145.644097753" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.003534 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:00 crc kubenswrapper[4799]: E1124 06:50:00.007624 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:00.507588627 +0000 UTC m=+146.163571261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.012967 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv" event={"ID":"fde84f5d-c045-4a70-9d7d-9f4ca71d4351","Type":"ContainerStarted","Data":"5df1347c0c3755ecc42eb4d10d291f363f5b1eea77417578b27a0f2e774cdaff"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.023973 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwzs" podStartSLOduration=123.023953729 podStartE2EDuration="2m3.023953729s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:00.02219943 +0000 UTC m=+145.678181904" watchObservedRunningTime="2025-11-24 06:50:00.023953729 +0000 UTC m=+145.679936203" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.056819 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" event={"ID":"fec9f210-0d28-463f-b3bf-4f27c6eee903","Type":"ContainerStarted","Data":"64bde5bee0f330aeeb08d6cdcbf32068ab2d01af213d9b46af66799e339d6405"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.073935 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" event={"ID":"892c2271-ba94-41b9-b5ed-04d45c2b1194","Type":"ContainerStarted","Data":"759049a215549d7a6d0d511da3bb4fdf67288acb5cbe679f3ac2220d89fdbf33"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.077578 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" event={"ID":"2927bbfb-073e-41b2-bb65-c6c75a54560b","Type":"ContainerStarted","Data":"efe583c357ff6f069bffc6981f9e9c20bda5d6356c3fc2614b2907b12d8726d2"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.078385 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fd7xr" podStartSLOduration=123.078345296 podStartE2EDuration="2m3.078345296s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:00.075259622 +0000 UTC m=+145.731242096" watchObservedRunningTime="2025-11-24 06:50:00.078345296 +0000 UTC m=+145.734327770" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.080006 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" event={"ID":"626e983a-0392-49ce-9b4b-96c2db7320eb","Type":"ContainerStarted","Data":"df37fefc945a28309311ef1b5c6da986d30e7fec5c42a3657028b035a8592e22"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.087755 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" event={"ID":"86f640cf-7c54-4b44-a52f-21378c535d5a","Type":"ContainerStarted","Data":"2a52c088ffcb045f85e46e6c4cf57f101dd1b1be68f5f805e5f9a4833c39529c"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.105533 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" event={"ID":"61518cfd-2e9e-4d2f-95ed-612afa636cef","Type":"ContainerStarted","Data":"e2cb3bfa4dca843e998aaabda30cd016b117e3bf8629ba11d7c0484478f9513f"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.106693 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:00 crc kubenswrapper[4799]: E1124 06:50:00.109605 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:00.609587412 +0000 UTC m=+146.265569886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.134178 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dwshf" podStartSLOduration=124.134147291 podStartE2EDuration="2m4.134147291s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:00.104520811 +0000 UTC m=+145.760503285" watchObservedRunningTime="2025-11-24 06:50:00.134147291 +0000 UTC m=+145.790129765" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.166567 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln" event={"ID":"5e8415c7-91a0-4d83-9627-d7e394393334","Type":"ContainerStarted","Data":"b53fa5361e0aaf310a81de75090204f157d8cb8ec76dc6ed09c2f30ca9310d97"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.212212 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6j6bn" event={"ID":"d6a12d20-1525-4cf5-8e5c-da7184c42581","Type":"ContainerStarted","Data":"2f8251d2be01216844db44a7059cfa859d6ffb3bdd1a4ddaeaf067cd9962da92"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.213515 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6j6bn" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.220410 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:00 crc kubenswrapper[4799]: E1124 06:50:00.220912 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:00.72086851 +0000 UTC m=+146.376850994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.237826 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hqzcz" podStartSLOduration=123.237805642 podStartE2EDuration="2m3.237805642s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:00.232162882 +0000 UTC m=+145.888145376" watchObservedRunningTime="2025-11-24 06:50:00.237805642 +0000 UTC m=+145.893788116" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.238296 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-rhhsg" podStartSLOduration=123.238292219 podStartE2EDuration="2m3.238292219s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:00.136793201 +0000 UTC m=+145.792775675" watchObservedRunningTime="2025-11-24 06:50:00.238292219 +0000 UTC m=+145.894274693" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.251352 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-6j6bn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.251431 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6j6bn" podUID="d6a12d20-1525-4cf5-8e5c-da7184c42581" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.251771 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-27fxx" event={"ID":"7edcb436-bf5b-4afd-9603-4a8131b2be16","Type":"ContainerStarted","Data":"d76050347815c17f990bddb478e7a73552ab5ba18431db1406cda32f10fd7153"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.266691 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" event={"ID":"17a805ea-4e7f-4204-9712-8be2c673363d","Type":"ContainerStarted","Data":"86322f61211036e853f9a4e3ff2dabd4897d05b52b9c05b7fc5efe657bccf9b2"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.303161 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" event={"ID":"22faf60d-16a2-49ae-9eb5-5f808013081e","Type":"ContainerStarted","Data":"675ae6a8ff45cc28b7dd061db3090f4c012bb2aa96d18b1c195c54ab4df4c1ab"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.329037 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:00 crc kubenswrapper[4799]: E1124 06:50:00.330560 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:00.830538245 +0000 UTC m=+146.486520719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.341505 4799 generic.go:334] "Generic (PLEG): container finished" podID="1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f" containerID="52f52697cafc64d25fd438f51fdf2826dfa09bd2bbc57193f16444dc3765660e" exitCode=0 Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.341596 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" event={"ID":"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f","Type":"ContainerDied","Data":"52f52697cafc64d25fd438f51fdf2826dfa09bd2bbc57193f16444dc3765660e"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.397408 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" event={"ID":"facf7448-3c40-476b-bc7f-9526c5f4c3e2","Type":"ContainerStarted","Data":"89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.398641 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.414606 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:00 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:00 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:00 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.414676 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.416626 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" event={"ID":"ed31a762-47f8-4efa-88f4-262e93cbdcea","Type":"ContainerStarted","Data":"cab877a5595618da2a1d11ed3a36ecc76bee5c92ed975de8ca586ae72f1975fd"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.427364 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-6j6bn" podStartSLOduration=123.427327724 podStartE2EDuration="2m3.427327724s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:00.309796974 +0000 UTC m=+145.965779448" watchObservedRunningTime="2025-11-24 06:50:00.427327724 +0000 UTC m=+146.083310188" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.428976 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" event={"ID":"c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf","Type":"ContainerStarted","Data":"41a47d8158bd3ac699e66cab6a8ca7eedaefabf645c0c17626e806d233765ddf"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.430720 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:00 crc kubenswrapper[4799]: E1124 06:50:00.431455 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:00.931433063 +0000 UTC m=+146.587415537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.455132 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" event={"ID":"7aede9e9-6b07-4959-8097-982b17d5628d","Type":"ContainerStarted","Data":"784d53818eeb25135d3e619388fd9fcd26eebfd6625ea05235b7af09b20757df"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.456484 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" podStartSLOduration=123.456467698 podStartE2EDuration="2m3.456467698s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:00.45591503 +0000 UTC m=+146.111897514" watchObservedRunningTime="2025-11-24 06:50:00.456467698 +0000 UTC m=+146.112450172" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.465573 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" event={"ID":"0625b922-f42b-4378-ba25-1ee3fc85cce6","Type":"ContainerStarted","Data":"dbdc6670e4e47ff59beb3bb39fbf0ca14d91db28637665077d4d2b1d39c53ed7"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.476801 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" event={"ID":"7dc41818-1f5e-4472-a726-f1322a1f462b","Type":"ContainerStarted","Data":"6b1e2cd271677e8f1dc88e56e97fa4b0dca01f62a727bb324436b5810a142a94"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.491622 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" event={"ID":"bfae07f4-0529-426f-aaac-d82964808045","Type":"ContainerStarted","Data":"67b41157d46732e600176e23f6e0ca07595cfb60d58196d0a2c331ab78a1b38f"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.500619 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" event={"ID":"3b02c876-c14b-4422-beaf-9ea3ab910f90","Type":"ContainerStarted","Data":"0455c604bc8a8efd5be8f2379d90380f01a095a7fef4d48bfc97e5e12f004f72"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.510990 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" event={"ID":"7f394c75-c5d2-4b2a-8845-d2337bc1735e","Type":"ContainerStarted","Data":"8946123ab04c96eeedb0eda8f4096b812fbf8e4a772ceee2d7c1264b0ebd1790"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.526237 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-dv8qs" event={"ID":"5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f","Type":"ContainerStarted","Data":"112b908fe7bd662fe423dfdeadddf6ce9dc7380034b828fb120f59c992df4be8"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.532398 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:00 crc kubenswrapper[4799]: E1124 06:50:00.537345 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:01.037320759 +0000 UTC m=+146.693303413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.545393 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" event={"ID":"7776e83d-dc11-4222-8440-cf059292e127","Type":"ContainerStarted","Data":"e40632a459eab330348cabffdd7e6a5e380a8b642a47490aa24353ffe57fd3b8"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.563513 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qv7f9" podStartSLOduration=123.563490953 podStartE2EDuration="2m3.563490953s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:00.556499207 +0000 UTC m=+146.212481681" watchObservedRunningTime="2025-11-24 06:50:00.563490953 +0000 UTC m=+146.219473427" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.563536 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" event={"ID":"4708bfd8-5610-47ff-a842-d2bc1c6e05bc","Type":"ContainerStarted","Data":"a1de659052ef58c12ab23bfddd549f6c698928c60bdb5b654a6317cce5bd1b08"} Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.571050 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" podStartSLOduration=123.571038458 podStartE2EDuration="2m3.571038458s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:00.521852257 +0000 UTC m=+146.177834721" watchObservedRunningTime="2025-11-24 06:50:00.571038458 +0000 UTC m=+146.227020932" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.572237 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-zzd4b" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.573766 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.641497 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:00 crc kubenswrapper[4799]: E1124 06:50:00.642283 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:01.142256114 +0000 UTC m=+146.798238588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.642967 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" podStartSLOduration=123.642946547 podStartE2EDuration="2m3.642946547s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:00.641059423 +0000 UTC m=+146.297041897" watchObservedRunningTime="2025-11-24 06:50:00.642946547 +0000 UTC m=+146.298929021" Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.772241 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:00 crc kubenswrapper[4799]: E1124 06:50:00.774537 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:01.274515221 +0000 UTC m=+146.930497695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.875821 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:00 crc kubenswrapper[4799]: E1124 06:50:00.876152 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:01.376110332 +0000 UTC m=+147.032092806 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.876611 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:00 crc kubenswrapper[4799]: E1124 06:50:00.877139 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:01.377120057 +0000 UTC m=+147.033102521 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:00 crc kubenswrapper[4799]: I1124 06:50:00.978228 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:00 crc kubenswrapper[4799]: E1124 06:50:00.978799 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:01.47876864 +0000 UTC m=+147.134751114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.062766 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.080481 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:01 crc kubenswrapper[4799]: E1124 06:50:01.080958 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:01.580942041 +0000 UTC m=+147.236924515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.185718 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:01 crc kubenswrapper[4799]: E1124 06:50:01.186124 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:01.686102583 +0000 UTC m=+147.342085057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.292055 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:01 crc kubenswrapper[4799]: E1124 06:50:01.294257 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:01.794239786 +0000 UTC m=+147.450222260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.399788 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:01 crc kubenswrapper[4799]: E1124 06:50:01.400284 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:01.900261997 +0000 UTC m=+147.556244471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.405804 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:01 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:01 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:01 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.405892 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.501324 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:01 crc kubenswrapper[4799]: E1124 06:50:01.502095 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:02.002080886 +0000 UTC m=+147.658063360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.584086 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2npgz" event={"ID":"0625b922-f42b-4378-ba25-1ee3fc85cce6","Type":"ContainerStarted","Data":"9d3c5aa081414f8d8c71b3166355525f65e8dc8f3fc7ac1cfbfd4e688ba9120e"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.602667 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" event={"ID":"baecc37e-a64f-449f-bde9-f4ff3a66e628","Type":"ContainerStarted","Data":"05995fd49856f8d98921634b86c0f981ef2af8b6fbbaa98edcd863683be1adf1"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.602725 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" event={"ID":"baecc37e-a64f-449f-bde9-f4ff3a66e628","Type":"ContainerStarted","Data":"c41d0bda5bd1998a198dad0fda682d775a4c082e87357016dfca9422ecb877ae"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.607543 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:01 crc kubenswrapper[4799]: E1124 06:50:01.607934 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:02.107917861 +0000 UTC m=+147.763900335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.681832 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-bbm7w" podStartSLOduration=124.681801036 podStartE2EDuration="2m4.681801036s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:01.669210211 +0000 UTC m=+147.325192685" watchObservedRunningTime="2025-11-24 06:50:01.681801036 +0000 UTC m=+147.337783510" Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.709284 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:01 crc kubenswrapper[4799]: E1124 06:50:01.711258 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:02.211242321 +0000 UTC m=+147.867224795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.814505 4799 generic.go:334] "Generic (PLEG): container finished" podID="22faf60d-16a2-49ae-9eb5-5f808013081e" containerID="97d2d23f7f2d15329954073b094af28c65f6a7c5b76c17fa24830c406e6af73f" exitCode=0 Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.815123 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" podStartSLOduration=124.815098048 podStartE2EDuration="2m4.815098048s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:01.724374114 +0000 UTC m=+147.380356588" watchObservedRunningTime="2025-11-24 06:50:01.815098048 +0000 UTC m=+147.471080522" Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.831461 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:01 crc kubenswrapper[4799]: E1124 06:50:01.831667 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:02.331632857 +0000 UTC m=+147.987615331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.832512 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:01 crc kubenswrapper[4799]: E1124 06:50:01.834360 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:02.334346189 +0000 UTC m=+147.990328663 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.862567 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" event={"ID":"17a805ea-4e7f-4204-9712-8be2c673363d","Type":"ContainerStarted","Data":"e1199f5fa16cfdb97de5e26f969db91bd37b9009edee8ea84ec853942a21549f"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.862632 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jf7t" event={"ID":"7f394c75-c5d2-4b2a-8845-d2337bc1735e","Type":"ContainerStarted","Data":"4ae40fadfb3f5b65572900050432841519c3b1e0c5e68031f1106c143c0495e0"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.862652 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" event={"ID":"86f640cf-7c54-4b44-a52f-21378c535d5a","Type":"ContainerStarted","Data":"247f5b993f2b5c52bcedf7867a9ce47d83a2010330da88c3ae493768a50f816c"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.862663 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" event={"ID":"61518cfd-2e9e-4d2f-95ed-612afa636cef","Type":"ContainerStarted","Data":"ba5ed616d5d5c354d004df33513c54c2aba76c6e666b782de58c50e0634020c6"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.862674 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv" event={"ID":"fde84f5d-c045-4a70-9d7d-9f4ca71d4351","Type":"ContainerStarted","Data":"ac6b9d57fa0b6b3b7798e84e6e084080f3557fbdfe9209742c54898466808bb2"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.862688 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-58mkm" event={"ID":"336a419e-e807-4859-b624-2ed06a9a8665","Type":"ContainerStarted","Data":"72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.862699 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" event={"ID":"4708bfd8-5610-47ff-a842-d2bc1c6e05bc","Type":"ContainerStarted","Data":"751bb0c17e49f96d91db320be81da66171fbb0a454d68e01fa5bfc9345a428b2"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.862710 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" event={"ID":"22faf60d-16a2-49ae-9eb5-5f808013081e","Type":"ContainerDied","Data":"97d2d23f7f2d15329954073b094af28c65f6a7c5b76c17fa24830c406e6af73f"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.862728 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-27fxx" event={"ID":"7edcb436-bf5b-4afd-9603-4a8131b2be16","Type":"ContainerStarted","Data":"cc07a1d5944bbc125416ba7720d50b04bf874f907f76d2b98e615e9a3587c1df"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.862739 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w4d6j" event={"ID":"66a7f09b-b285-4635-8666-b8caaa05b1c7","Type":"ContainerStarted","Data":"c9fb60fbb3d37e6bab6a205a38c7855dec1d246b3ab1f5e3a2ce001b4f6ffb3e"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.879356 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" event={"ID":"9663e76d-9ba9-496c-a654-7493c2aa1268","Type":"ContainerStarted","Data":"622a3187ae41f577a939d79925ff172f8e89db8a3c312deb9641aaeb9fad574b"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.880137 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.919538 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" event={"ID":"3b02c876-c14b-4422-beaf-9ea3ab910f90","Type":"ContainerStarted","Data":"b4c7b3100a3fc2990475a72d5d9965c9288ff6e1f0c3c62513f32bcf494ab724"} Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.919572 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6hw2" podStartSLOduration=124.919549757 podStartE2EDuration="2m4.919549757s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:01.81782264 +0000 UTC m=+147.473805114" watchObservedRunningTime="2025-11-24 06:50:01.919549757 +0000 UTC m=+147.575532231" Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.920246 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.920758 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-knplv" podStartSLOduration=124.920751567 podStartE2EDuration="2m4.920751567s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:01.919463964 +0000 UTC m=+147.575446438" watchObservedRunningTime="2025-11-24 06:50:01.920751567 +0000 UTC m=+147.576734041" Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.923974 4799 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-t4wbk container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.924021 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" podUID="9663e76d-9ba9-496c-a654-7493c2aa1268" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.935057 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.936509 4799 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-pcvtw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Nov 24 06:50:01 crc kubenswrapper[4799]: E1124 06:50:01.936586 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:02.436561891 +0000 UTC m=+148.092544355 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:01 crc kubenswrapper[4799]: I1124 06:50:01.936595 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" podUID="3b02c876-c14b-4422-beaf-9ea3ab910f90" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:01.942301 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" event={"ID":"fec9f210-0d28-463f-b3bf-4f27c6eee903","Type":"ContainerStarted","Data":"43a419309c30bd82156efdd692d5ec703cd39c7b27349a6074f98919639d9089"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:01.958674 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" event={"ID":"892c2271-ba94-41b9-b5ed-04d45c2b1194","Type":"ContainerStarted","Data":"dfe2e4f343f38fe198d68179222bfbbde8541ceac69c88c37e6b63da4a8eaec1"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:01.973468 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" event={"ID":"7776e83d-dc11-4222-8440-cf059292e127","Type":"ContainerStarted","Data":"80fb8d10184fc6c7c66e8f557e785c5ba2d85fbc10d9b7a8f45efa3453175df8"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:01.975921 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" event={"ID":"626e983a-0392-49ce-9b4b-96c2db7320eb","Type":"ContainerStarted","Data":"8798f01d939b2ef2f3fa2237d3500978e6ac4de70e6faf4d790118b83c9edee7"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:01.992410 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-58mkm" podStartSLOduration=125.992383337 podStartE2EDuration="2m5.992383337s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:01.991473376 +0000 UTC m=+147.647455850" watchObservedRunningTime="2025-11-24 06:50:01.992383337 +0000 UTC m=+147.648365811" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.036945 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.040292 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:02.540277574 +0000 UTC m=+148.196260058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.055530 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-64b4w" podStartSLOduration=125.055503259 podStartE2EDuration="2m5.055503259s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.05525004 +0000 UTC m=+147.711232514" watchObservedRunningTime="2025-11-24 06:50:02.055503259 +0000 UTC m=+147.711485733" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.124229 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln" event={"ID":"5e8415c7-91a0-4d83-9627-d7e394393334","Type":"ContainerStarted","Data":"f707a15468a11f4c78e72f1b7d6f20bf62c1324fd9af668de8b4d81ee24ce60c"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.142938 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.143294 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" podStartSLOduration=125.143258623 podStartE2EDuration="2m5.143258623s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.091178824 +0000 UTC m=+147.747161318" watchObservedRunningTime="2025-11-24 06:50:02.143258623 +0000 UTC m=+147.799241097" Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.144027 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:02.644006398 +0000 UTC m=+148.299988872 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.145486 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-nqlvh" podStartSLOduration=125.145467787 podStartE2EDuration="2m5.145467787s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.141952179 +0000 UTC m=+147.797934663" watchObservedRunningTime="2025-11-24 06:50:02.145467787 +0000 UTC m=+147.801450271" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.156245 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" event={"ID":"9b717be8-b446-40c0-a5d9-1de2525168fd","Type":"ContainerStarted","Data":"42234c1f7e9dbd0f7ba9407279e1aa8bc6a145911252412a9b4c2842e34e3d93"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.203529 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" event={"ID":"7aede9e9-6b07-4959-8097-982b17d5628d","Type":"ContainerStarted","Data":"d9278b58575654d747b5dd107a084a9ff27ce75d0d7e7e4e0810d37ec359773d"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.204781 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.218260 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" event={"ID":"9e51884d-3817-4148-8556-5d91843c2e91","Type":"ContainerStarted","Data":"9b70f7f8ebab3af29497e766bef2309adf186dabc18675d91a34cf449c8a560d"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.220560 4799 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6d9nh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.220625 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" podUID="7aede9e9-6b07-4959-8097-982b17d5628d" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.237098 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-27fxx" podStartSLOduration=8.237074482 podStartE2EDuration="8.237074482s" podCreationTimestamp="2025-11-24 06:49:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.227498188 +0000 UTC m=+147.883480662" watchObservedRunningTime="2025-11-24 06:50:02.237074482 +0000 UTC m=+147.893056956" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.246283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" event={"ID":"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af","Type":"ContainerStarted","Data":"0a24edf3800235ddc274a2510c46ed1b231361918ada2e6609f1b56b091c9c73"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.246350 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" event={"ID":"a3fa15d5-b246-4af2-9fbc-55bc4d5de8af","Type":"ContainerStarted","Data":"247943c2222c6238a334ee331a89331754eb0c95ed1494a4dd46134094bc536c"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.247517 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.248639 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.253221 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:02.753198076 +0000 UTC m=+148.409180740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.264531 4799 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-qn45g container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.264733 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" podUID="a3fa15d5-b246-4af2-9fbc-55bc4d5de8af" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.265468 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" event={"ID":"e116addf-dac0-42aa-9c0e-82383b214399","Type":"ContainerStarted","Data":"94819cffa015d52177e06ae6f5f57848a48242776101bfdf308311b381b8e2e3"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.295295 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" event={"ID":"7dc41818-1f5e-4472-a726-f1322a1f462b","Type":"ContainerStarted","Data":"fb4f1c9a2efff87d19c91ec02c301953d04f734c9102e857df83174952a28059"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.296311 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.302271 4799 generic.go:334] "Generic (PLEG): container finished" podID="bfae07f4-0529-426f-aaac-d82964808045" containerID="67b41157d46732e600176e23f6e0ca07595cfb60d58196d0a2c331ab78a1b38f" exitCode=0 Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.302395 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" event={"ID":"bfae07f4-0529-426f-aaac-d82964808045","Type":"ContainerDied","Data":"67b41157d46732e600176e23f6e0ca07595cfb60d58196d0a2c331ab78a1b38f"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.313040 4799 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-7gmz9 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.36:6443/healthz\": dial tcp 10.217.0.36:6443: connect: connection refused" start-of-body= Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.313128 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" podUID="7dc41818-1f5e-4472-a726-f1322a1f462b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.36:6443/healthz\": dial tcp 10.217.0.36:6443: connect: connection refused" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.347512 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" podStartSLOduration=125.347489481 podStartE2EDuration="2m5.347489481s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.276031287 +0000 UTC m=+147.932013761" watchObservedRunningTime="2025-11-24 06:50:02.347489481 +0000 UTC m=+148.003471945" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.349216 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njqfl" podStartSLOduration=126.349208429 podStartE2EDuration="2m6.349208429s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.346323972 +0000 UTC m=+148.002306446" watchObservedRunningTime="2025-11-24 06:50:02.349208429 +0000 UTC m=+148.005190903" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.350446 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.354312 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-dv8qs" event={"ID":"5b91c061-9cd7-41e8-b37b-9ed80f9b1c3f","Type":"ContainerStarted","Data":"38ec08a2a49aef2a11e451f625fef5cc9b797a5809be496eb26743819ca78d66"} Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.355682 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:02.855657457 +0000 UTC m=+148.511640111 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.371630 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" podStartSLOduration=125.371602965 podStartE2EDuration="2m5.371602965s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.371354167 +0000 UTC m=+148.027336641" watchObservedRunningTime="2025-11-24 06:50:02.371602965 +0000 UTC m=+148.027585439" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.391907 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nf9tt" event={"ID":"c8bb6a9e-70e5-4a09-a1df-e94eceb02ebf","Type":"ContainerStarted","Data":"172f2d2b2b37c6e391a3fe29b6b59ca9d6a92a6249b065d5f5e02250d34c80f0"} Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.398047 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-6j6bn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.398112 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6j6bn" podUID="d6a12d20-1525-4cf5-8e5c-da7184c42581" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.419175 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:02 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:02 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:02 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.419250 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.421916 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" podStartSLOduration=126.421899084 podStartE2EDuration="2m6.421899084s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.421427998 +0000 UTC m=+148.077410472" watchObservedRunningTime="2025-11-24 06:50:02.421899084 +0000 UTC m=+148.077881558" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.457927 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.458383 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:02.958362346 +0000 UTC m=+148.614344820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.467967 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" podStartSLOduration=125.46794405 podStartE2EDuration="2m5.46794405s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.463786969 +0000 UTC m=+148.119769443" watchObservedRunningTime="2025-11-24 06:50:02.46794405 +0000 UTC m=+148.123926524" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.568992 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.569819 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.069780939 +0000 UTC m=+148.725763413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.570225 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.556524 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" podStartSLOduration=125.556500761 podStartE2EDuration="2m5.556500761s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.51326493 +0000 UTC m=+148.169247404" watchObservedRunningTime="2025-11-24 06:50:02.556500761 +0000 UTC m=+148.212483235" Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.574314 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" podStartSLOduration=125.574293062 podStartE2EDuration="2m5.574293062s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.555445325 +0000 UTC m=+148.211427799" watchObservedRunningTime="2025-11-24 06:50:02.574293062 +0000 UTC m=+148.230275536" Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.582376 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.082356004 +0000 UTC m=+148.738338478 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.671480 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.672438 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.172415606 +0000 UTC m=+148.828398080 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.778946 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.779399 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.279381619 +0000 UTC m=+148.935364093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.880324 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.880587 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.380547776 +0000 UTC m=+149.036530250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.881275 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.881762 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.381751717 +0000 UTC m=+149.037734191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.983265 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.983432 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.48340022 +0000 UTC m=+149.139382694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:02 crc kubenswrapper[4799]: I1124 06:50:02.983935 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:02 crc kubenswrapper[4799]: E1124 06:50:02.984394 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.484385213 +0000 UTC m=+149.140367687 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.085617 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.085832 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.585796589 +0000 UTC m=+149.241779063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.086038 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.086486 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.586476462 +0000 UTC m=+149.242459116 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.187683 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.187937 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.687898807 +0000 UTC m=+149.343881281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.188202 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.188706 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.688674224 +0000 UTC m=+149.344656698 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.289841 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.290022 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.789989986 +0000 UTC m=+149.445972460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.290171 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.290611 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.790596076 +0000 UTC m=+149.446578540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.390913 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.391135 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.891104711 +0000 UTC m=+149.547087185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.391367 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.391752 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.891732332 +0000 UTC m=+149.547714806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.429133 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:03 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:03 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:03 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.429214 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.440718 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d9jlc" event={"ID":"9e51884d-3817-4148-8556-5d91843c2e91","Type":"ContainerStarted","Data":"7e869fa3dc87695f8d9461c6bca5148ae8f5101e6fb225fbb2c51bdefd76d8b3"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.449452 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln" event={"ID":"5e8415c7-91a0-4d83-9627-d7e394393334","Type":"ContainerStarted","Data":"c4a413d2028a2984120b3c34209340a7afa44f5721f0d52482b191d06fc1969c"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.459365 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" event={"ID":"17a805ea-4e7f-4204-9712-8be2c673363d","Type":"ContainerStarted","Data":"0f2365718a9ca75aa2ad8af15282229ed62c8804f1827ba54a5df8231fb473cd"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.468943 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" event={"ID":"22faf60d-16a2-49ae-9eb5-5f808013081e","Type":"ContainerStarted","Data":"b7f5540ff44e27e5a7cf76fc6968d48401620c6724983f21e6229ffdbc76f9b1"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.495045 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.497401 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:03.99737759 +0000 UTC m=+149.653360064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.497560 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" event={"ID":"4708bfd8-5610-47ff-a842-d2bc1c6e05bc","Type":"ContainerStarted","Data":"f3b48aae59dfab2554e0ba2053d7b194136d033434e66439ffebbe2e04bb41a3"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.516115 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-dv8qs" podStartSLOduration=8.516080891 podStartE2EDuration="8.516080891s" podCreationTimestamp="2025-11-24 06:49:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:02.604180481 +0000 UTC m=+148.260162955" watchObservedRunningTime="2025-11-24 06:50:03.516080891 +0000 UTC m=+149.172063365" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.531276 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" event={"ID":"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f","Type":"ContainerStarted","Data":"f7b7a87e52a0f0b348cf4be65dbcf363f567dd6b92fa7403a9e5ef70d3c96781"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.531342 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" event={"ID":"1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f","Type":"ContainerStarted","Data":"0133944047dc101302e810cbd0497aa5341599cdfebe4183e641012a5df1ac10"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.569574 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" event={"ID":"bfae07f4-0529-426f-aaac-d82964808045","Type":"ContainerStarted","Data":"2fd0df4511bb7c87baaf6cc22fb7ac07628cf2332212afb99caa9262df1292f7"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.570482 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.589075 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w4d6j" event={"ID":"66a7f09b-b285-4635-8666-b8caaa05b1c7","Type":"ContainerStarted","Data":"5eedd110c6c644e47e2e19cd3766b57dc475061ae780bbc2c452ec02e913314a"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.589125 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w4d6j" event={"ID":"66a7f09b-b285-4635-8666-b8caaa05b1c7","Type":"ContainerStarted","Data":"8670a4629b41ab7e1c06db3308afa7c5a74ace33951021ced5d4b7ab60705c88"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.589956 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-w4d6j" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.594462 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pxzr4" podStartSLOduration=126.594449438 podStartE2EDuration="2m6.594449438s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:03.592363508 +0000 UTC m=+149.248345982" watchObservedRunningTime="2025-11-24 06:50:03.594449438 +0000 UTC m=+149.250431912" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.595165 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79kln" podStartSLOduration=126.595160082 podStartE2EDuration="2m6.595160082s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:03.518979089 +0000 UTC m=+149.174961563" watchObservedRunningTime="2025-11-24 06:50:03.595160082 +0000 UTC m=+149.251142546" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.597076 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.597122 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.597343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.597372 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.597406 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.600689 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:04.100668858 +0000 UTC m=+149.756651332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.603150 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.606797 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.637599 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.638158 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.653075 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" podStartSLOduration=126.653031447 podStartE2EDuration="2m6.653031447s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:03.64837742 +0000 UTC m=+149.304359894" watchObservedRunningTime="2025-11-24 06:50:03.653031447 +0000 UTC m=+149.309013921" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.655470 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" event={"ID":"892c2271-ba94-41b9-b5ed-04d45c2b1194","Type":"ContainerStarted","Data":"195e78782b146d0e5c7970dcbb6ad25d69c1d2228f149fde201b629492fc29e4"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.655519 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" event={"ID":"2927bbfb-073e-41b2-bb65-c6c75a54560b","Type":"ContainerStarted","Data":"24e37bb6fcfb5d3639b88da945d00b5254364cec027df52d607d2747a5ed0289"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.672389 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.687286 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" event={"ID":"7776e83d-dc11-4222-8440-cf059292e127","Type":"ContainerStarted","Data":"e9a427a59bf71ee465c61d329314e022cdeadd2bc292dc9864d455e286471de2"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.688485 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.690597 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.702377 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.703331 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:04.203307045 +0000 UTC m=+149.859289519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.707642 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.757134 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" event={"ID":"61518cfd-2e9e-4d2f-95ed-612afa636cef","Type":"ContainerStarted","Data":"e3e9a15098ab8ea57fade9b631c25c34f1275b2037cae92303a46065f75df5d1"} Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.757337 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-6j6bn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.757443 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6j6bn" podUID="d6a12d20-1525-4cf5-8e5c-da7184c42581" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.757629 4799 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-pcvtw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.757742 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" podUID="3b02c876-c14b-4422-beaf-9ea3ab910f90" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.781834 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-t4wbk" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.789240 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d9nh" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.804362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.815716 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-n2sth" podStartSLOduration=126.815691081 podStartE2EDuration="2m6.815691081s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:03.701352199 +0000 UTC m=+149.357334693" watchObservedRunningTime="2025-11-24 06:50:03.815691081 +0000 UTC m=+149.471673555" Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.817672 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:04.317654977 +0000 UTC m=+149.973637451 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.846539 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.898599 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" podStartSLOduration=126.898575891 podStartE2EDuration="2m6.898575891s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:03.816956314 +0000 UTC m=+149.472938788" watchObservedRunningTime="2025-11-24 06:50:03.898575891 +0000 UTC m=+149.554558365" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.898835 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" podStartSLOduration=126.898830539 podStartE2EDuration="2m6.898830539s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:03.879421454 +0000 UTC m=+149.535403928" watchObservedRunningTime="2025-11-24 06:50:03.898830539 +0000 UTC m=+149.554813003" Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.905736 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.906315 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:04.406286281 +0000 UTC m=+150.062268755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.909687 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:03 crc kubenswrapper[4799]: E1124 06:50:03.910111 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:04.4100946 +0000 UTC m=+150.066077074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:03 crc kubenswrapper[4799]: I1124 06:50:03.923092 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9wp9k" podStartSLOduration=126.923065628 podStartE2EDuration="2m6.923065628s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:03.921318659 +0000 UTC m=+149.577301133" watchObservedRunningTime="2025-11-24 06:50:03.923065628 +0000 UTC m=+149.579048102" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.004742 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-w4d6j" podStartSLOduration=9.004715556 podStartE2EDuration="9.004715556s" podCreationTimestamp="2025-11-24 06:49:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:03.964541929 +0000 UTC m=+149.620524403" watchObservedRunningTime="2025-11-24 06:50:04.004715556 +0000 UTC m=+149.660698030" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.006296 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" podStartSLOduration=128.006290249 podStartE2EDuration="2m8.006290249s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:04.003414942 +0000 UTC m=+149.659397416" watchObservedRunningTime="2025-11-24 06:50:04.006290249 +0000 UTC m=+149.662272723" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.014442 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.014799 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:04.514784736 +0000 UTC m=+150.170767210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.075973 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4h5t" podStartSLOduration=127.075955762 podStartE2EDuration="2m7.075955762s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:04.074796663 +0000 UTC m=+149.730779137" watchObservedRunningTime="2025-11-24 06:50:04.075955762 +0000 UTC m=+149.731938236" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.117625 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.118198 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:04.618183828 +0000 UTC m=+150.274166302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.151650 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r2hbg"] Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.170923 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.191415 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r2hbg"] Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.198768 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.223691 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.224088 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrk6v\" (UniqueName: \"kubernetes.io/projected/787fe453-1786-449c-bf60-87153ce058f9-kube-api-access-rrk6v\") pod \"community-operators-r2hbg\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.224132 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-catalog-content\") pod \"community-operators-r2hbg\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.224185 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-utilities\") pod \"community-operators-r2hbg\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.224320 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:04.724298843 +0000 UTC m=+150.380281317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.235765 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-csh4j"] Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.236916 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.253776 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.329274 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-utilities\") pod \"community-operators-r2hbg\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.330455 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmdxc\" (UniqueName: \"kubernetes.io/projected/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-kube-api-access-nmdxc\") pod \"certified-operators-csh4j\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.333023 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-utilities\") pod \"certified-operators-csh4j\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.341082 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.330549 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-utilities\") pod \"community-operators-r2hbg\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.342346 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:04.842323699 +0000 UTC m=+150.498306173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.343747 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrk6v\" (UniqueName: \"kubernetes.io/projected/787fe453-1786-449c-bf60-87153ce058f9-kube-api-access-rrk6v\") pod \"community-operators-r2hbg\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.344147 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-catalog-content\") pod \"community-operators-r2hbg\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.348699 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-catalog-content\") pod \"certified-operators-csh4j\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.347526 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-catalog-content\") pod \"community-operators-r2hbg\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.354946 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-csh4j"] Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.392128 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrk6v\" (UniqueName: \"kubernetes.io/projected/787fe453-1786-449c-bf60-87153ce058f9-kube-api-access-rrk6v\") pod \"community-operators-r2hbg\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.421944 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:04 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:04 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:04 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.422250 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.431486 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sglfw"] Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.432599 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.454223 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.454610 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-catalog-content\") pod \"certified-operators-csh4j\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.454730 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmdxc\" (UniqueName: \"kubernetes.io/projected/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-kube-api-access-nmdxc\") pod \"certified-operators-csh4j\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.454823 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-utilities\") pod \"certified-operators-csh4j\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.455405 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-utilities\") pod \"certified-operators-csh4j\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.455554 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:04.955530603 +0000 UTC m=+150.611513067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.473570 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sglfw"] Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.477265 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qn45g" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.502220 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmdxc\" (UniqueName: \"kubernetes.io/projected/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-kube-api-access-nmdxc\") pod \"certified-operators-csh4j\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.557509 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-catalog-content\") pod \"certified-operators-csh4j\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.558349 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-utilities\") pod \"community-operators-sglfw\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.558379 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-catalog-content\") pod \"community-operators-sglfw\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.558407 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6fl2\" (UniqueName: \"kubernetes.io/projected/ea43b679-a4d2-4539-bcf3-130fbfe86656-kube-api-access-f6fl2\") pod \"community-operators-sglfw\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.558453 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.558807 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.058793701 +0000 UTC m=+150.714776175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.576069 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.629996 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pvkkn"] Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.631346 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.635279 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.664787 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pvkkn"] Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.677592 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.677930 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-utilities\") pod \"community-operators-sglfw\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.678436 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-utilities\") pod \"community-operators-sglfw\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.678554 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.178528235 +0000 UTC m=+150.834510709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.678579 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-catalog-content\") pod \"community-operators-sglfw\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.678618 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6fl2\" (UniqueName: \"kubernetes.io/projected/ea43b679-a4d2-4539-bcf3-130fbfe86656-kube-api-access-f6fl2\") pod \"community-operators-sglfw\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.678675 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.679085 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.179075573 +0000 UTC m=+150.835058057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.679518 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-catalog-content\") pod \"community-operators-sglfw\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.737797 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6fl2\" (UniqueName: \"kubernetes.io/projected/ea43b679-a4d2-4539-bcf3-130fbfe86656-kube-api-access-f6fl2\") pod \"community-operators-sglfw\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.781632 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.782110 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.282076322 +0000 UTC m=+150.938058796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.782270 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.782728 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-utilities\") pod \"certified-operators-pvkkn\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.782780 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.782840 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6xks\" (UniqueName: \"kubernetes.io/projected/04184df2-64d2-4175-a1fd-a83670d7df6c-kube-api-access-n6xks\") pod \"certified-operators-pvkkn\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.782913 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-catalog-content\") pod \"certified-operators-pvkkn\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.783420 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.283407727 +0000 UTC m=+150.939390201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.807213 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" event={"ID":"2927bbfb-073e-41b2-bb65-c6c75a54560b","Type":"ContainerStarted","Data":"c6e68f2a9666460adba0aa332ed6f6fb71cf9a46c1675f78918d5b27e0875932"} Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.810139 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"245caa35bfc194d7e24f8c3592edaeaf27e2437a223e8078c43977d9766169ff"} Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.885662 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.886545 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6xks\" (UniqueName: \"kubernetes.io/projected/04184df2-64d2-4175-a1fd-a83670d7df6c-kube-api-access-n6xks\") pod \"certified-operators-pvkkn\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.886654 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.386594123 +0000 UTC m=+151.042576607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.887109 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-catalog-content\") pod \"certified-operators-pvkkn\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.887522 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-utilities\") pod \"certified-operators-pvkkn\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.887852 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.888820 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-utilities\") pod \"certified-operators-pvkkn\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.892087 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-catalog-content\") pod \"certified-operators-pvkkn\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:50:04 crc kubenswrapper[4799]: E1124 06:50:04.893948 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.393931391 +0000 UTC m=+151.049913865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:04 crc kubenswrapper[4799]: I1124 06:50:04.921002 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6xks\" (UniqueName: \"kubernetes.io/projected/04184df2-64d2-4175-a1fd-a83670d7df6c-kube-api-access-n6xks\") pod \"certified-operators-pvkkn\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:50:04 crc kubenswrapper[4799]: W1124 06:50:04.936284 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-8398ef85a8e7f37cef4bc1b06aa76fc19cb35d5ab1b3d98be142e65638766960 WatchSource:0}: Error finding container 8398ef85a8e7f37cef4bc1b06aa76fc19cb35d5ab1b3d98be142e65638766960: Status 404 returned error can't find the container with id 8398ef85a8e7f37cef4bc1b06aa76fc19cb35d5ab1b3d98be142e65638766960 Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:04.999046 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:05 crc kubenswrapper[4799]: E1124 06:50:04.999998 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.499968112 +0000 UTC m=+151.155950586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.100988 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:05 crc kubenswrapper[4799]: E1124 06:50:05.101643 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.601627066 +0000 UTC m=+151.257609540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.120404 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.131826 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r2hbg"] Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.202560 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:05 crc kubenswrapper[4799]: E1124 06:50:05.203037 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.703015681 +0000 UTC m=+151.358998155 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.304176 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:05 crc kubenswrapper[4799]: E1124 06:50:05.305434 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.805418589 +0000 UTC m=+151.461401063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.405822 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:05 crc kubenswrapper[4799]: E1124 06:50:05.406290 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:05.906268826 +0000 UTC m=+151.562251300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.409609 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:05 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:05 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:05 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.409687 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.436411 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-csh4j"] Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.451950 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sglfw"] Nov 24 06:50:05 crc kubenswrapper[4799]: W1124 06:50:05.474196 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea43b679_a4d2_4539_bcf3_130fbfe86656.slice/crio-cd009638c8dde9be47a5f963ff3690b94816838f7c213fd8a036b3b68393b87b WatchSource:0}: Error finding container cd009638c8dde9be47a5f963ff3690b94816838f7c213fd8a036b3b68393b87b: Status 404 returned error can't find the container with id cd009638c8dde9be47a5f963ff3690b94816838f7c213fd8a036b3b68393b87b Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.507542 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:05 crc kubenswrapper[4799]: E1124 06:50:05.508014 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:06.007996282 +0000 UTC m=+151.663978756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.553901 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pvkkn"] Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.590076 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.591081 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.598825 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.600328 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.606486 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.610226 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:05 crc kubenswrapper[4799]: E1124 06:50:05.610791 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:06.110757443 +0000 UTC m=+151.766739917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:05 crc kubenswrapper[4799]: W1124 06:50:05.614587 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04184df2_64d2_4175_a1fd_a83670d7df6c.slice/crio-48e0ecedb19e609fceb32bbc15575b37771947a3b082376dcc434c6a4dd0c8c2 WatchSource:0}: Error finding container 48e0ecedb19e609fceb32bbc15575b37771947a3b082376dcc434c6a4dd0c8c2: Status 404 returned error can't find the container with id 48e0ecedb19e609fceb32bbc15575b37771947a3b082376dcc434c6a4dd0c8c2 Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.712158 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed153767-6876-4cb0-9344-cfc0133c9ff8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ed153767-6876-4cb0-9344-cfc0133c9ff8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.712203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed153767-6876-4cb0-9344-cfc0133c9ff8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ed153767-6876-4cb0-9344-cfc0133c9ff8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.712455 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:05 crc kubenswrapper[4799]: E1124 06:50:05.713001 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:06.212984556 +0000 UTC m=+151.868967030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.814587 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.815315 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed153767-6876-4cb0-9344-cfc0133c9ff8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ed153767-6876-4cb0-9344-cfc0133c9ff8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.815355 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed153767-6876-4cb0-9344-cfc0133c9ff8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ed153767-6876-4cb0-9344-cfc0133c9ff8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:50:05 crc kubenswrapper[4799]: E1124 06:50:05.815465 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:06.315427986 +0000 UTC m=+151.971410460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.815673 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed153767-6876-4cb0-9344-cfc0133c9ff8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ed153767-6876-4cb0-9344-cfc0133c9ff8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.828276 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2hbg" event={"ID":"787fe453-1786-449c-bf60-87153ce058f9","Type":"ContainerStarted","Data":"f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a"} Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.828339 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2hbg" event={"ID":"787fe453-1786-449c-bf60-87153ce058f9","Type":"ContainerStarted","Data":"8b1f0c47b9a118f38140e805eca6a66aef24bfe127cf597c3e144878bad78472"} Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.833094 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" event={"ID":"2927bbfb-073e-41b2-bb65-c6c75a54560b","Type":"ContainerStarted","Data":"c566e54e69534dbc2bc66d795245c27744f1f714b98eb5eea1c65f9d7113640a"} Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.838662 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"32a2ce58bc220639f9fa7b30b807ad2067dba862b6f2f39931d36fe608f9315a"} Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.838870 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.844800 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ec3133675054c77420bbd0ee47532d0a130791edbe800dc886f0100644d223d5"} Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.844846 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8398ef85a8e7f37cef4bc1b06aa76fc19cb35d5ab1b3d98be142e65638766960"} Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.850043 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csh4j" event={"ID":"3476fe21-9207-4c63-80e9-0c3e1bd7e19c","Type":"ContainerStarted","Data":"74f6e98ea94738701935e11fa028e308b4ec56cbbb363cba2159eda110dc0d1e"} Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.852364 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"140be9742174f3bff62ccc7c6ef9f7c237009525a33a3540a0e98c27b256c244"} Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.852421 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"f783031a5d3e0f42547ff9f793bf06e928a7907db6022dd876e1f858f5051517"} Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.857807 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sglfw" event={"ID":"ea43b679-a4d2-4539-bcf3-130fbfe86656","Type":"ContainerStarted","Data":"cd009638c8dde9be47a5f963ff3690b94816838f7c213fd8a036b3b68393b87b"} Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.870752 4799 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.882730 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvkkn" event={"ID":"04184df2-64d2-4175-a1fd-a83670d7df6c","Type":"ContainerStarted","Data":"48e0ecedb19e609fceb32bbc15575b37771947a3b082376dcc434c6a4dd0c8c2"} Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.900091 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed153767-6876-4cb0-9344-cfc0133c9ff8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ed153767-6876-4cb0-9344-cfc0133c9ff8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.917302 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-glxff" Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.917979 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:05 crc kubenswrapper[4799]: E1124 06:50:05.918362 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 06:50:06.418339832 +0000 UTC m=+152.074322306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5bjhn" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:05 crc kubenswrapper[4799]: I1124 06:50:05.981635 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.004619 4799 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T06:50:05.871058315Z","Handler":null,"Name":""} Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.018774 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:06 crc kubenswrapper[4799]: E1124 06:50:06.020937 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 06:50:06.520907816 +0000 UTC m=+152.176890290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.025085 4799 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.025128 4799 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.120878 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.139748 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.139798 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.204247 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5bjhn\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.222629 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.223798 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-288hz"] Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.237888 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-288hz"] Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.238096 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.243420 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.245203 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.280030 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 06:50:06 crc kubenswrapper[4799]: W1124 06:50:06.317215 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poded153767_6876_4cb0_9344_cfc0133c9ff8.slice/crio-502aabe583961642fcd1c4fd2bf7031e6067892e6496350e61c2e59c6d6ba742 WatchSource:0}: Error finding container 502aabe583961642fcd1c4fd2bf7031e6067892e6496350e61c2e59c6d6ba742: Status 404 returned error can't find the container with id 502aabe583961642fcd1c4fd2bf7031e6067892e6496350e61c2e59c6d6ba742 Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.324502 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-catalog-content\") pod \"redhat-marketplace-288hz\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.324608 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-utilities\") pod \"redhat-marketplace-288hz\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.324663 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvx26\" (UniqueName: \"kubernetes.io/projected/31145cc8-41cc-4870-83f8-bca24095bb0c-kube-api-access-fvx26\") pod \"redhat-marketplace-288hz\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.349779 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.406072 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:06 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:06 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:06 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.406141 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.426388 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-utilities\") pod \"redhat-marketplace-288hz\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.426448 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvx26\" (UniqueName: \"kubernetes.io/projected/31145cc8-41cc-4870-83f8-bca24095bb0c-kube-api-access-fvx26\") pod \"redhat-marketplace-288hz\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.426528 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-catalog-content\") pod \"redhat-marketplace-288hz\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.427043 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-catalog-content\") pod \"redhat-marketplace-288hz\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.427280 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-utilities\") pod \"redhat-marketplace-288hz\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.452473 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvx26\" (UniqueName: \"kubernetes.io/projected/31145cc8-41cc-4870-83f8-bca24095bb0c-kube-api-access-fvx26\") pod \"redhat-marketplace-288hz\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.622768 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5bjhn"] Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.623217 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-htg5n"] Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.624796 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.631439 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.638040 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-htg5n"] Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.735872 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-catalog-content\") pod \"redhat-marketplace-htg5n\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.736290 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-utilities\") pod \"redhat-marketplace-htg5n\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.736434 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8ffn\" (UniqueName: \"kubernetes.io/projected/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-kube-api-access-v8ffn\") pod \"redhat-marketplace-htg5n\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.838090 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8ffn\" (UniqueName: \"kubernetes.io/projected/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-kube-api-access-v8ffn\") pod \"redhat-marketplace-htg5n\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.838211 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-catalog-content\") pod \"redhat-marketplace-htg5n\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.838229 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-utilities\") pod \"redhat-marketplace-htg5n\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.838794 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-utilities\") pod \"redhat-marketplace-htg5n\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.839371 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-catalog-content\") pod \"redhat-marketplace-htg5n\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.872044 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8ffn\" (UniqueName: \"kubernetes.io/projected/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-kube-api-access-v8ffn\") pod \"redhat-marketplace-htg5n\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.890014 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" event={"ID":"b7761980-8663-4dda-924b-2fb787fcdac8","Type":"ContainerStarted","Data":"c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1"} Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.890078 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" event={"ID":"b7761980-8663-4dda-924b-2fb787fcdac8","Type":"ContainerStarted","Data":"78e78ea44ec9153887432e0a140338778ae9d10337a0c152ec7319ba64863063"} Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.891068 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.892360 4799 generic.go:334] "Generic (PLEG): container finished" podID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" containerID="73b29289760a17b5c99878355b59e3c29d1c7de4850786852fb28091cb8875ce" exitCode=0 Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.892573 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csh4j" event={"ID":"3476fe21-9207-4c63-80e9-0c3e1bd7e19c","Type":"ContainerDied","Data":"73b29289760a17b5c99878355b59e3c29d1c7de4850786852fb28091cb8875ce"} Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.898431 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.899523 4799 generic.go:334] "Generic (PLEG): container finished" podID="ea43b679-a4d2-4539-bcf3-130fbfe86656" containerID="f21d29de5814bec4cedc52e0f1441c6750e24186bb2ef8663700a14f765c3bf9" exitCode=0 Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.899648 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sglfw" event={"ID":"ea43b679-a4d2-4539-bcf3-130fbfe86656","Type":"ContainerDied","Data":"f21d29de5814bec4cedc52e0f1441c6750e24186bb2ef8663700a14f765c3bf9"} Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.908434 4799 generic.go:334] "Generic (PLEG): container finished" podID="04184df2-64d2-4175-a1fd-a83670d7df6c" containerID="8539cae693510d551953b7ebacd79e759935d9c7a486993d561a55ad7e584a18" exitCode=0 Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.909802 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvkkn" event={"ID":"04184df2-64d2-4175-a1fd-a83670d7df6c","Type":"ContainerDied","Data":"8539cae693510d551953b7ebacd79e759935d9c7a486993d561a55ad7e584a18"} Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.911586 4799 generic.go:334] "Generic (PLEG): container finished" podID="787fe453-1786-449c-bf60-87153ce058f9" containerID="f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a" exitCode=0 Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.911637 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2hbg" event={"ID":"787fe453-1786-449c-bf60-87153ce058f9","Type":"ContainerDied","Data":"f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a"} Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.919249 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" podStartSLOduration=129.917029484 podStartE2EDuration="2m9.917029484s" podCreationTimestamp="2025-11-24 06:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:06.916268309 +0000 UTC m=+152.572250813" watchObservedRunningTime="2025-11-24 06:50:06.917029484 +0000 UTC m=+152.573011958" Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.923208 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" event={"ID":"2927bbfb-073e-41b2-bb65-c6c75a54560b","Type":"ContainerStarted","Data":"2e55a5c1c52a422f2fec2810ac1684ce1454705248c5212e1676f6ff210fde1f"} Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.929013 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ed153767-6876-4cb0-9344-cfc0133c9ff8","Type":"ContainerStarted","Data":"e562893ac9b3067dff79419a731e0e2cafbad3c7969593c8185bbb1d0d66f0cb"} Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.929086 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ed153767-6876-4cb0-9344-cfc0133c9ff8","Type":"ContainerStarted","Data":"502aabe583961642fcd1c4fd2bf7031e6067892e6496350e61c2e59c6d6ba742"} Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.948254 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-288hz"] Nov 24 06:50:06 crc kubenswrapper[4799]: I1124 06:50:06.998269 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.069975 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-2zg6b" podStartSLOduration=13.069945728 podStartE2EDuration="13.069945728s" podCreationTimestamp="2025-11-24 06:49:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:07.067164914 +0000 UTC m=+152.723147388" watchObservedRunningTime="2025-11-24 06:50:07.069945728 +0000 UTC m=+152.725928212" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.097028 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.097002762 podStartE2EDuration="2.097002762s" podCreationTimestamp="2025-11-24 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:07.096176275 +0000 UTC m=+152.752158749" watchObservedRunningTime="2025-11-24 06:50:07.097002762 +0000 UTC m=+152.752985236" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.233226 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vzb6m"] Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.246130 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.246176 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.246189 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vzb6m"] Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.247202 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.252083 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.360410 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-utilities\") pod \"redhat-operators-vzb6m\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.360467 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-catalog-content\") pod \"redhat-operators-vzb6m\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.360617 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ftkp\" (UniqueName: \"kubernetes.io/projected/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-kube-api-access-9ftkp\") pod \"redhat-operators-vzb6m\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.374445 4799 patch_prober.go:28] interesting pod/apiserver-76f77b778f-b28hc container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 24 06:50:07 crc kubenswrapper[4799]: [+]log ok Nov 24 06:50:07 crc kubenswrapper[4799]: [+]etcd ok Nov 24 06:50:07 crc kubenswrapper[4799]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 24 06:50:07 crc kubenswrapper[4799]: [+]poststarthook/generic-apiserver-start-informers ok Nov 24 06:50:07 crc kubenswrapper[4799]: [+]poststarthook/max-in-flight-filter ok Nov 24 06:50:07 crc kubenswrapper[4799]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 24 06:50:07 crc kubenswrapper[4799]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 24 06:50:07 crc kubenswrapper[4799]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 24 06:50:07 crc kubenswrapper[4799]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 24 06:50:07 crc kubenswrapper[4799]: [+]poststarthook/project.openshift.io-projectcache ok Nov 24 06:50:07 crc kubenswrapper[4799]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 24 06:50:07 crc kubenswrapper[4799]: [+]poststarthook/openshift.io-startinformers ok Nov 24 06:50:07 crc kubenswrapper[4799]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 24 06:50:07 crc kubenswrapper[4799]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 24 06:50:07 crc kubenswrapper[4799]: livez check failed Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.374564 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" podUID="1ca0d9c6-5437-4ef7-bca4-f7d5d1b33d7f" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.382147 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.382233 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.384580 4799 patch_prober.go:28] interesting pod/console-f9d7485db-58mkm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.35:8443/health\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.384679 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-58mkm" podUID="336a419e-e807-4859-b624-2ed06a9a8665" containerName="console" probeResult="failure" output="Get \"https://10.217.0.35:8443/health\": dial tcp 10.217.0.35:8443: connect: connection refused" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.401018 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.407657 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:07 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:07 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:07 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.407732 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.461886 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-utilities\") pod \"redhat-operators-vzb6m\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.461949 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-catalog-content\") pod \"redhat-operators-vzb6m\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.461993 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ftkp\" (UniqueName: \"kubernetes.io/projected/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-kube-api-access-9ftkp\") pod \"redhat-operators-vzb6m\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.463361 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-utilities\") pod \"redhat-operators-vzb6m\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.463670 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-catalog-content\") pod \"redhat-operators-vzb6m\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.489118 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ftkp\" (UniqueName: \"kubernetes.io/projected/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-kube-api-access-9ftkp\") pod \"redhat-operators-vzb6m\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.491949 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-htg5n"] Nov 24 06:50:07 crc kubenswrapper[4799]: W1124 06:50:07.500083 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0cc7caf_7a38_4cd6_a809_098d38b8cf3a.slice/crio-bb003b979b928fd26af0774325783a3f319efd3f7dffcf748f33080a52cf8b25 WatchSource:0}: Error finding container bb003b979b928fd26af0774325783a3f319efd3f7dffcf748f33080a52cf8b25: Status 404 returned error can't find the container with id bb003b979b928fd26af0774325783a3f319efd3f7dffcf748f33080a52cf8b25 Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.594682 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-6j6bn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.594767 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6j6bn" podUID="d6a12d20-1525-4cf5-8e5c-da7184c42581" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.594929 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.595365 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-6j6bn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.595394 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6j6bn" podUID="d6a12d20-1525-4cf5-8e5c-da7184c42581" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.622261 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rqx5j"] Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.625733 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.660347 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.661177 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rqx5j"] Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.774190 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-utilities\") pod \"redhat-operators-rqx5j\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.774684 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-catalog-content\") pod \"redhat-operators-rqx5j\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.774801 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n994\" (UniqueName: \"kubernetes.io/projected/e6461490-0f33-4730-83a1-7d01ac7f863b-kube-api-access-4n994\") pod \"redhat-operators-rqx5j\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.876285 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-catalog-content\") pod \"redhat-operators-rqx5j\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.876407 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n994\" (UniqueName: \"kubernetes.io/projected/e6461490-0f33-4730-83a1-7d01ac7f863b-kube-api-access-4n994\") pod \"redhat-operators-rqx5j\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.876491 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-utilities\") pod \"redhat-operators-rqx5j\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.877354 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-utilities\") pod \"redhat-operators-rqx5j\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.877681 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-catalog-content\") pod \"redhat-operators-rqx5j\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.885369 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vzb6m"] Nov 24 06:50:07 crc kubenswrapper[4799]: W1124 06:50:07.898262 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dab8ddb_4cb5_4349_a4e5_22b50fbf0740.slice/crio-2f4b1ecb25ed5a425c4f859749db0c70af0d41e666003b74fafd0a73d84bce2d WatchSource:0}: Error finding container 2f4b1ecb25ed5a425c4f859749db0c70af0d41e666003b74fafd0a73d84bce2d: Status 404 returned error can't find the container with id 2f4b1ecb25ed5a425c4f859749db0c70af0d41e666003b74fafd0a73d84bce2d Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.912233 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n994\" (UniqueName: \"kubernetes.io/projected/e6461490-0f33-4730-83a1-7d01ac7f863b-kube-api-access-4n994\") pod \"redhat-operators-rqx5j\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.973129 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.973990 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.980708 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzb6m" event={"ID":"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740","Type":"ContainerStarted","Data":"2f4b1ecb25ed5a425c4f859749db0c70af0d41e666003b74fafd0a73d84bce2d"} Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.984173 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.987807 4799 generic.go:334] "Generic (PLEG): container finished" podID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" containerID="fda22a58e5759c18bb559846e28eb163132efbd27783e06e3600e9a0f8692be5" exitCode=0 Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.987920 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htg5n" event={"ID":"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a","Type":"ContainerDied","Data":"fda22a58e5759c18bb559846e28eb163132efbd27783e06e3600e9a0f8692be5"} Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.987960 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htg5n" event={"ID":"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a","Type":"ContainerStarted","Data":"bb003b979b928fd26af0774325783a3f319efd3f7dffcf748f33080a52cf8b25"} Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.992235 4799 generic.go:334] "Generic (PLEG): container finished" podID="e116addf-dac0-42aa-9c0e-82383b214399" containerID="94819cffa015d52177e06ae6f5f57848a48242776101bfdf308311b381b8e2e3" exitCode=0 Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.992347 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" event={"ID":"e116addf-dac0-42aa-9c0e-82383b214399","Type":"ContainerDied","Data":"94819cffa015d52177e06ae6f5f57848a48242776101bfdf308311b381b8e2e3"} Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.993133 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.996804 4799 generic.go:334] "Generic (PLEG): container finished" podID="ed153767-6876-4cb0-9344-cfc0133c9ff8" containerID="e562893ac9b3067dff79419a731e0e2cafbad3c7969593c8185bbb1d0d66f0cb" exitCode=0 Nov 24 06:50:07 crc kubenswrapper[4799]: I1124 06:50:07.996907 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ed153767-6876-4cb0-9344-cfc0133c9ff8","Type":"ContainerDied","Data":"e562893ac9b3067dff79419a731e0e2cafbad3c7969593c8185bbb1d0d66f0cb"} Nov 24 06:50:08 crc kubenswrapper[4799]: I1124 06:50:08.013794 4799 generic.go:334] "Generic (PLEG): container finished" podID="31145cc8-41cc-4870-83f8-bca24095bb0c" containerID="a85acdb86cfd13022e9cd801757268441a2382f918ccca53f464209c2696ed81" exitCode=0 Nov 24 06:50:08 crc kubenswrapper[4799]: I1124 06:50:08.014987 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-288hz" event={"ID":"31145cc8-41cc-4870-83f8-bca24095bb0c","Type":"ContainerDied","Data":"a85acdb86cfd13022e9cd801757268441a2382f918ccca53f464209c2696ed81"} Nov 24 06:50:08 crc kubenswrapper[4799]: I1124 06:50:08.015038 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-288hz" event={"ID":"31145cc8-41cc-4870-83f8-bca24095bb0c","Type":"ContainerStarted","Data":"9e233586edb956dfb147e59a6bb27b1fed603bca262d06844378f9f3bb6376e2"} Nov 24 06:50:08 crc kubenswrapper[4799]: I1124 06:50:08.127015 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:50:08 crc kubenswrapper[4799]: I1124 06:50:08.347694 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rqx5j"] Nov 24 06:50:08 crc kubenswrapper[4799]: W1124 06:50:08.387268 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6461490_0f33_4730_83a1_7d01ac7f863b.slice/crio-eff00acff6928b0481371d3e1ab6445debe7cfcb3f4fc865f7b86ff066e39ad2 WatchSource:0}: Error finding container eff00acff6928b0481371d3e1ab6445debe7cfcb3f4fc865f7b86ff066e39ad2: Status 404 returned error can't find the container with id eff00acff6928b0481371d3e1ab6445debe7cfcb3f4fc865f7b86ff066e39ad2 Nov 24 06:50:08 crc kubenswrapper[4799]: I1124 06:50:08.405026 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:08 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:08 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:08 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:08 crc kubenswrapper[4799]: I1124 06:50:08.405089 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.030461 4799 generic.go:334] "Generic (PLEG): container finished" podID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerID="e7c2a2fbcce322aff186b83de04e2832885951194c8be888842552b7d0b5a8d4" exitCode=0 Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.030775 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rqx5j" event={"ID":"e6461490-0f33-4730-83a1-7d01ac7f863b","Type":"ContainerDied","Data":"e7c2a2fbcce322aff186b83de04e2832885951194c8be888842552b7d0b5a8d4"} Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.031234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rqx5j" event={"ID":"e6461490-0f33-4730-83a1-7d01ac7f863b","Type":"ContainerStarted","Data":"eff00acff6928b0481371d3e1ab6445debe7cfcb3f4fc865f7b86ff066e39ad2"} Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.037720 4799 generic.go:334] "Generic (PLEG): container finished" podID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerID="ed401757ec819c88507bc7c4f79a0d833d180236d8e239ca36b691815549f8ae" exitCode=0 Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.038060 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzb6m" event={"ID":"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740","Type":"ContainerDied","Data":"ed401757ec819c88507bc7c4f79a0d833d180236d8e239ca36b691815549f8ae"} Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.044428 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wsxhz" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.270648 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.271741 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.274485 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.282745 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.283385 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.308152 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.410086 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63282372-2d7d-40fe-b64a-adea4d6931e3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"63282372-2d7d-40fe-b64a-adea4d6931e3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.410163 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63282372-2d7d-40fe-b64a-adea4d6931e3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"63282372-2d7d-40fe-b64a-adea4d6931e3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.439421 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:09 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:09 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:09 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.439481 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.469611 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.511742 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed153767-6876-4cb0-9344-cfc0133c9ff8-kubelet-dir\") pod \"ed153767-6876-4cb0-9344-cfc0133c9ff8\" (UID: \"ed153767-6876-4cb0-9344-cfc0133c9ff8\") " Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.511905 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed153767-6876-4cb0-9344-cfc0133c9ff8-kube-api-access\") pod \"ed153767-6876-4cb0-9344-cfc0133c9ff8\" (UID: \"ed153767-6876-4cb0-9344-cfc0133c9ff8\") " Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.512276 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63282372-2d7d-40fe-b64a-adea4d6931e3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"63282372-2d7d-40fe-b64a-adea4d6931e3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.512308 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63282372-2d7d-40fe-b64a-adea4d6931e3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"63282372-2d7d-40fe-b64a-adea4d6931e3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.512763 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed153767-6876-4cb0-9344-cfc0133c9ff8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ed153767-6876-4cb0-9344-cfc0133c9ff8" (UID: "ed153767-6876-4cb0-9344-cfc0133c9ff8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.513455 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63282372-2d7d-40fe-b64a-adea4d6931e3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"63282372-2d7d-40fe-b64a-adea4d6931e3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.522602 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed153767-6876-4cb0-9344-cfc0133c9ff8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ed153767-6876-4cb0-9344-cfc0133c9ff8" (UID: "ed153767-6876-4cb0-9344-cfc0133c9ff8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.534096 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.538381 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63282372-2d7d-40fe-b64a-adea4d6931e3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"63282372-2d7d-40fe-b64a-adea4d6931e3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.608637 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.613198 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e116addf-dac0-42aa-9c0e-82383b214399-config-volume\") pod \"e116addf-dac0-42aa-9c0e-82383b214399\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.613921 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e116addf-dac0-42aa-9c0e-82383b214399-secret-volume\") pod \"e116addf-dac0-42aa-9c0e-82383b214399\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.613973 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e116addf-dac0-42aa-9c0e-82383b214399-config-volume" (OuterVolumeSpecName: "config-volume") pod "e116addf-dac0-42aa-9c0e-82383b214399" (UID: "e116addf-dac0-42aa-9c0e-82383b214399"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.614093 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwdqk\" (UniqueName: \"kubernetes.io/projected/e116addf-dac0-42aa-9c0e-82383b214399-kube-api-access-mwdqk\") pod \"e116addf-dac0-42aa-9c0e-82383b214399\" (UID: \"e116addf-dac0-42aa-9c0e-82383b214399\") " Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.614971 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e116addf-dac0-42aa-9c0e-82383b214399-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.614993 4799 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed153767-6876-4cb0-9344-cfc0133c9ff8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.615005 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed153767-6876-4cb0-9344-cfc0133c9ff8-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.618155 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e116addf-dac0-42aa-9c0e-82383b214399-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e116addf-dac0-42aa-9c0e-82383b214399" (UID: "e116addf-dac0-42aa-9c0e-82383b214399"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.618411 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e116addf-dac0-42aa-9c0e-82383b214399-kube-api-access-mwdqk" (OuterVolumeSpecName: "kube-api-access-mwdqk") pod "e116addf-dac0-42aa-9c0e-82383b214399" (UID: "e116addf-dac0-42aa-9c0e-82383b214399"). InnerVolumeSpecName "kube-api-access-mwdqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.716587 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e116addf-dac0-42aa-9c0e-82383b214399-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.716620 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwdqk\" (UniqueName: \"kubernetes.io/projected/e116addf-dac0-42aa-9c0e-82383b214399-kube-api-access-mwdqk\") on node \"crc\" DevicePath \"\"" Nov 24 06:50:09 crc kubenswrapper[4799]: I1124 06:50:09.934294 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 06:50:10 crc kubenswrapper[4799]: I1124 06:50:10.057520 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" Nov 24 06:50:10 crc kubenswrapper[4799]: I1124 06:50:10.057556 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd" event={"ID":"e116addf-dac0-42aa-9c0e-82383b214399","Type":"ContainerDied","Data":"462aeb8d0ee119d1da5c9d2e27b1fd83726d314733294f99511ab1672114f0ab"} Nov 24 06:50:10 crc kubenswrapper[4799]: I1124 06:50:10.058313 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="462aeb8d0ee119d1da5c9d2e27b1fd83726d314733294f99511ab1672114f0ab" Nov 24 06:50:10 crc kubenswrapper[4799]: I1124 06:50:10.061458 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"63282372-2d7d-40fe-b64a-adea4d6931e3","Type":"ContainerStarted","Data":"c722a48b891fb46f6a6df18c6c765a0c3202c10ff7b085fc955c3cbf6da05972"} Nov 24 06:50:10 crc kubenswrapper[4799]: I1124 06:50:10.088016 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ed153767-6876-4cb0-9344-cfc0133c9ff8","Type":"ContainerDied","Data":"502aabe583961642fcd1c4fd2bf7031e6067892e6496350e61c2e59c6d6ba742"} Nov 24 06:50:10 crc kubenswrapper[4799]: I1124 06:50:10.088111 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="502aabe583961642fcd1c4fd2bf7031e6067892e6496350e61c2e59c6d6ba742" Nov 24 06:50:10 crc kubenswrapper[4799]: I1124 06:50:10.088047 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 06:50:10 crc kubenswrapper[4799]: I1124 06:50:10.403521 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:10 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:10 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:10 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:10 crc kubenswrapper[4799]: I1124 06:50:10.403631 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:11 crc kubenswrapper[4799]: I1124 06:50:11.118399 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"63282372-2d7d-40fe-b64a-adea4d6931e3","Type":"ContainerStarted","Data":"ae9df859f9b8693fd9028b3c6bf37f1d2951701c1752dd2c0f2b95caa2cc8e52"} Nov 24 06:50:11 crc kubenswrapper[4799]: I1124 06:50:11.403428 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:11 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:11 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:11 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:11 crc kubenswrapper[4799]: I1124 06:50:11.405834 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:12 crc kubenswrapper[4799]: I1124 06:50:12.136683 4799 generic.go:334] "Generic (PLEG): container finished" podID="63282372-2d7d-40fe-b64a-adea4d6931e3" containerID="ae9df859f9b8693fd9028b3c6bf37f1d2951701c1752dd2c0f2b95caa2cc8e52" exitCode=0 Nov 24 06:50:12 crc kubenswrapper[4799]: I1124 06:50:12.136756 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"63282372-2d7d-40fe-b64a-adea4d6931e3","Type":"ContainerDied","Data":"ae9df859f9b8693fd9028b3c6bf37f1d2951701c1752dd2c0f2b95caa2cc8e52"} Nov 24 06:50:12 crc kubenswrapper[4799]: I1124 06:50:12.242522 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:50:12 crc kubenswrapper[4799]: I1124 06:50:12.249344 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-b28hc" Nov 24 06:50:12 crc kubenswrapper[4799]: I1124 06:50:12.403805 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:12 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:12 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:12 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:12 crc kubenswrapper[4799]: I1124 06:50:12.403938 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:13 crc kubenswrapper[4799]: I1124 06:50:13.206894 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-w4d6j" Nov 24 06:50:13 crc kubenswrapper[4799]: I1124 06:50:13.404170 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:13 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:13 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:13 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:13 crc kubenswrapper[4799]: I1124 06:50:13.404248 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:14 crc kubenswrapper[4799]: I1124 06:50:14.402886 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:14 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:14 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:14 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:14 crc kubenswrapper[4799]: I1124 06:50:14.403215 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:15 crc kubenswrapper[4799]: I1124 06:50:15.402617 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:15 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:15 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:15 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:15 crc kubenswrapper[4799]: I1124 06:50:15.402707 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:16 crc kubenswrapper[4799]: I1124 06:50:16.402614 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:16 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Nov 24 06:50:16 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:16 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:16 crc kubenswrapper[4799]: I1124 06:50:16.402712 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:17 crc kubenswrapper[4799]: I1124 06:50:17.379804 4799 patch_prober.go:28] interesting pod/console-f9d7485db-58mkm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.35:8443/health\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Nov 24 06:50:17 crc kubenswrapper[4799]: I1124 06:50:17.380415 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-58mkm" podUID="336a419e-e807-4859-b624-2ed06a9a8665" containerName="console" probeResult="failure" output="Get \"https://10.217.0.35:8443/health\": dial tcp 10.217.0.35:8443: connect: connection refused" Nov 24 06:50:17 crc kubenswrapper[4799]: I1124 06:50:17.405236 4799 patch_prober.go:28] interesting pod/router-default-5444994796-cn7ch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 06:50:17 crc kubenswrapper[4799]: [+]has-synced ok Nov 24 06:50:17 crc kubenswrapper[4799]: [+]process-running ok Nov 24 06:50:17 crc kubenswrapper[4799]: healthz check failed Nov 24 06:50:17 crc kubenswrapper[4799]: I1124 06:50:17.405324 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7ch" podUID="1317991a-3615-4dc5-9cbe-232e04180a7c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 06:50:17 crc kubenswrapper[4799]: I1124 06:50:17.595380 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-6j6bn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 06:50:17 crc kubenswrapper[4799]: I1124 06:50:17.595475 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6j6bn" podUID="d6a12d20-1525-4cf5-8e5c-da7184c42581" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 06:50:17 crc kubenswrapper[4799]: I1124 06:50:17.595927 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-6j6bn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 24 06:50:17 crc kubenswrapper[4799]: I1124 06:50:17.596007 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6j6bn" podUID="d6a12d20-1525-4cf5-8e5c-da7184c42581" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 24 06:50:18 crc kubenswrapper[4799]: I1124 06:50:18.403302 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:50:18 crc kubenswrapper[4799]: I1124 06:50:18.405927 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-cn7ch" Nov 24 06:50:18 crc kubenswrapper[4799]: I1124 06:50:18.708783 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:50:18 crc kubenswrapper[4799]: I1124 06:50:18.777787 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63282372-2d7d-40fe-b64a-adea4d6931e3-kubelet-dir\") pod \"63282372-2d7d-40fe-b64a-adea4d6931e3\" (UID: \"63282372-2d7d-40fe-b64a-adea4d6931e3\") " Nov 24 06:50:18 crc kubenswrapper[4799]: I1124 06:50:18.777949 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63282372-2d7d-40fe-b64a-adea4d6931e3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "63282372-2d7d-40fe-b64a-adea4d6931e3" (UID: "63282372-2d7d-40fe-b64a-adea4d6931e3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:50:18 crc kubenswrapper[4799]: I1124 06:50:18.778316 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63282372-2d7d-40fe-b64a-adea4d6931e3-kube-api-access\") pod \"63282372-2d7d-40fe-b64a-adea4d6931e3\" (UID: \"63282372-2d7d-40fe-b64a-adea4d6931e3\") " Nov 24 06:50:18 crc kubenswrapper[4799]: I1124 06:50:18.780610 4799 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63282372-2d7d-40fe-b64a-adea4d6931e3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 06:50:18 crc kubenswrapper[4799]: I1124 06:50:18.787425 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63282372-2d7d-40fe-b64a-adea4d6931e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "63282372-2d7d-40fe-b64a-adea4d6931e3" (UID: "63282372-2d7d-40fe-b64a-adea4d6931e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:50:18 crc kubenswrapper[4799]: I1124 06:50:18.881728 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63282372-2d7d-40fe-b64a-adea4d6931e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 06:50:19 crc kubenswrapper[4799]: I1124 06:50:19.186737 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:50:19 crc kubenswrapper[4799]: I1124 06:50:19.196264 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f587ff07-479b-4c95-95cb-406faffcfb68-metrics-certs\") pod \"network-metrics-daemon-ldbr4\" (UID: \"f587ff07-479b-4c95-95cb-406faffcfb68\") " pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:50:19 crc kubenswrapper[4799]: I1124 06:50:19.214971 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"63282372-2d7d-40fe-b64a-adea4d6931e3","Type":"ContainerDied","Data":"c722a48b891fb46f6a6df18c6c765a0c3202c10ff7b085fc955c3cbf6da05972"} Nov 24 06:50:19 crc kubenswrapper[4799]: I1124 06:50:19.215070 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c722a48b891fb46f6a6df18c6c765a0c3202c10ff7b085fc955c3cbf6da05972" Nov 24 06:50:19 crc kubenswrapper[4799]: I1124 06:50:19.214997 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 06:50:19 crc kubenswrapper[4799]: I1124 06:50:19.255406 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ldbr4" Nov 24 06:50:20 crc kubenswrapper[4799]: I1124 06:50:20.401476 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:50:20 crc kubenswrapper[4799]: I1124 06:50:20.401569 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:50:26 crc kubenswrapper[4799]: I1124 06:50:26.357691 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:50:27 crc kubenswrapper[4799]: I1124 06:50:27.390022 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:50:27 crc kubenswrapper[4799]: I1124 06:50:27.393164 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:50:27 crc kubenswrapper[4799]: I1124 06:50:27.600571 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-6j6bn" Nov 24 06:50:33 crc kubenswrapper[4799]: I1124 06:50:33.624108 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ldbr4"] Nov 24 06:50:38 crc kubenswrapper[4799]: I1124 06:50:38.137371 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2pchl" Nov 24 06:50:44 crc kubenswrapper[4799]: I1124 06:50:44.032746 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 06:50:45 crc kubenswrapper[4799]: E1124 06:50:45.211425 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 06:50:45 crc kubenswrapper[4799]: E1124 06:50:45.212294 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nmdxc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-csh4j_openshift-marketplace(3476fe21-9207-4c63-80e9-0c3e1bd7e19c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 06:50:45 crc kubenswrapper[4799]: E1124 06:50:45.213783 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-csh4j" podUID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" Nov 24 06:50:47 crc kubenswrapper[4799]: E1124 06:50:47.426967 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-csh4j" podUID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" Nov 24 06:50:47 crc kubenswrapper[4799]: I1124 06:50:47.460544 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" event={"ID":"f587ff07-479b-4c95-95cb-406faffcfb68","Type":"ContainerStarted","Data":"ee2e2cad92a7f3521f57b19ac54c6adb8bb40a5fd6d7547739e1475a09c21cc8"} Nov 24 06:50:47 crc kubenswrapper[4799]: E1124 06:50:47.532316 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 24 06:50:47 crc kubenswrapper[4799]: E1124 06:50:47.532581 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4n994,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rqx5j_openshift-marketplace(e6461490-0f33-4730-83a1-7d01ac7f863b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 06:50:47 crc kubenswrapper[4799]: E1124 06:50:47.533906 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rqx5j" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" Nov 24 06:50:47 crc kubenswrapper[4799]: E1124 06:50:47.575196 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 06:50:47 crc kubenswrapper[4799]: E1124 06:50:47.575910 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n6xks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-pvkkn_openshift-marketplace(04184df2-64d2-4175-a1fd-a83670d7df6c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 06:50:47 crc kubenswrapper[4799]: E1124 06:50:47.577294 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-pvkkn" podUID="04184df2-64d2-4175-a1fd-a83670d7df6c" Nov 24 06:50:48 crc kubenswrapper[4799]: E1124 06:50:48.514292 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-pvkkn" podUID="04184df2-64d2-4175-a1fd-a83670d7df6c" Nov 24 06:50:48 crc kubenswrapper[4799]: E1124 06:50:48.514466 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rqx5j" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" Nov 24 06:50:48 crc kubenswrapper[4799]: E1124 06:50:48.621073 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 24 06:50:48 crc kubenswrapper[4799]: E1124 06:50:48.621289 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v8ffn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-htg5n_openshift-marketplace(e0cc7caf-7a38-4cd6-a809-098d38b8cf3a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 06:50:48 crc kubenswrapper[4799]: E1124 06:50:48.622529 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-htg5n" podUID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" Nov 24 06:50:48 crc kubenswrapper[4799]: E1124 06:50:48.723603 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 24 06:50:48 crc kubenswrapper[4799]: E1124 06:50:48.723816 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fvx26,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-288hz_openshift-marketplace(31145cc8-41cc-4870-83f8-bca24095bb0c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 06:50:48 crc kubenswrapper[4799]: E1124 06:50:48.725108 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-288hz" podUID="31145cc8-41cc-4870-83f8-bca24095bb0c" Nov 24 06:50:49 crc kubenswrapper[4799]: I1124 06:50:49.475038 4799 generic.go:334] "Generic (PLEG): container finished" podID="ea43b679-a4d2-4539-bcf3-130fbfe86656" containerID="a8b369f081f419cd8e3afc97c9706d4679d30073243402601df5922ca1d450eb" exitCode=0 Nov 24 06:50:49 crc kubenswrapper[4799]: I1124 06:50:49.475155 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sglfw" event={"ID":"ea43b679-a4d2-4539-bcf3-130fbfe86656","Type":"ContainerDied","Data":"a8b369f081f419cd8e3afc97c9706d4679d30073243402601df5922ca1d450eb"} Nov 24 06:50:49 crc kubenswrapper[4799]: I1124 06:50:49.477373 4799 generic.go:334] "Generic (PLEG): container finished" podID="787fe453-1786-449c-bf60-87153ce058f9" containerID="1914af0f9a86a50d913a5f0c25653b449863cc18ee45f1b368cccb7044a7a06e" exitCode=0 Nov 24 06:50:49 crc kubenswrapper[4799]: I1124 06:50:49.477507 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2hbg" event={"ID":"787fe453-1786-449c-bf60-87153ce058f9","Type":"ContainerDied","Data":"1914af0f9a86a50d913a5f0c25653b449863cc18ee45f1b368cccb7044a7a06e"} Nov 24 06:50:49 crc kubenswrapper[4799]: I1124 06:50:49.490813 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" event={"ID":"f587ff07-479b-4c95-95cb-406faffcfb68","Type":"ContainerStarted","Data":"638f02faeee1258e9490501925a692f7d38d2d3067aa301d2922537327104ece"} Nov 24 06:50:49 crc kubenswrapper[4799]: I1124 06:50:49.510094 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzb6m" event={"ID":"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740","Type":"ContainerStarted","Data":"d641cf589fd5a7e578892fa32d1277d6791180735fc1c1b55f40eefe15ca2f03"} Nov 24 06:50:49 crc kubenswrapper[4799]: E1124 06:50:49.512602 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-288hz" podUID="31145cc8-41cc-4870-83f8-bca24095bb0c" Nov 24 06:50:49 crc kubenswrapper[4799]: E1124 06:50:49.513170 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-htg5n" podUID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" Nov 24 06:50:50 crc kubenswrapper[4799]: I1124 06:50:50.401293 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:50:50 crc kubenswrapper[4799]: I1124 06:50:50.401839 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:50:50 crc kubenswrapper[4799]: I1124 06:50:50.518693 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ldbr4" event={"ID":"f587ff07-479b-4c95-95cb-406faffcfb68","Type":"ContainerStarted","Data":"12c5a1314dbb96175e5b0c00752ba7c5d9c8f0f7c7ead40eca561392a2148044"} Nov 24 06:50:50 crc kubenswrapper[4799]: I1124 06:50:50.521039 4799 generic.go:334] "Generic (PLEG): container finished" podID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerID="d641cf589fd5a7e578892fa32d1277d6791180735fc1c1b55f40eefe15ca2f03" exitCode=0 Nov 24 06:50:50 crc kubenswrapper[4799]: I1124 06:50:50.521102 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzb6m" event={"ID":"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740","Type":"ContainerDied","Data":"d641cf589fd5a7e578892fa32d1277d6791180735fc1c1b55f40eefe15ca2f03"} Nov 24 06:50:50 crc kubenswrapper[4799]: I1124 06:50:50.566759 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-ldbr4" podStartSLOduration=174.566733718 podStartE2EDuration="2m54.566733718s" podCreationTimestamp="2025-11-24 06:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:50:50.542046824 +0000 UTC m=+196.198029318" watchObservedRunningTime="2025-11-24 06:50:50.566733718 +0000 UTC m=+196.222716202" Nov 24 06:50:51 crc kubenswrapper[4799]: I1124 06:50:51.533025 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sglfw" event={"ID":"ea43b679-a4d2-4539-bcf3-130fbfe86656","Type":"ContainerStarted","Data":"53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5"} Nov 24 06:50:51 crc kubenswrapper[4799]: I1124 06:50:51.537444 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2hbg" event={"ID":"787fe453-1786-449c-bf60-87153ce058f9","Type":"ContainerStarted","Data":"51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a"} Nov 24 06:50:51 crc kubenswrapper[4799]: I1124 06:50:51.568892 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sglfw" podStartSLOduration=4.036331401 podStartE2EDuration="47.568825115s" podCreationTimestamp="2025-11-24 06:50:04 +0000 UTC" firstStartedPulling="2025-11-24 06:50:06.903984104 +0000 UTC m=+152.559966568" lastFinishedPulling="2025-11-24 06:50:50.436477808 +0000 UTC m=+196.092460282" observedRunningTime="2025-11-24 06:50:51.561352963 +0000 UTC m=+197.217335457" watchObservedRunningTime="2025-11-24 06:50:51.568825115 +0000 UTC m=+197.224807609" Nov 24 06:50:51 crc kubenswrapper[4799]: I1124 06:50:51.586033 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r2hbg" podStartSLOduration=3.822610323 podStartE2EDuration="47.586008706s" podCreationTimestamp="2025-11-24 06:50:04 +0000 UTC" firstStartedPulling="2025-11-24 06:50:06.920594605 +0000 UTC m=+152.576577079" lastFinishedPulling="2025-11-24 06:50:50.683992978 +0000 UTC m=+196.339975462" observedRunningTime="2025-11-24 06:50:51.58464812 +0000 UTC m=+197.240630624" watchObservedRunningTime="2025-11-24 06:50:51.586008706 +0000 UTC m=+197.241991180" Nov 24 06:50:52 crc kubenswrapper[4799]: I1124 06:50:52.552687 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzb6m" event={"ID":"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740","Type":"ContainerStarted","Data":"ecf3144ce85fdf6e45be5912761a06823a3a942e244d8b1dd07d2dff1b728aff"} Nov 24 06:50:52 crc kubenswrapper[4799]: I1124 06:50:52.589413 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vzb6m" podStartSLOduration=3.095706385 podStartE2EDuration="45.589377352s" podCreationTimestamp="2025-11-24 06:50:07 +0000 UTC" firstStartedPulling="2025-11-24 06:50:09.04380771 +0000 UTC m=+154.699790184" lastFinishedPulling="2025-11-24 06:50:51.537478677 +0000 UTC m=+197.193461151" observedRunningTime="2025-11-24 06:50:52.58558133 +0000 UTC m=+198.241563854" watchObservedRunningTime="2025-11-24 06:50:52.589377352 +0000 UTC m=+198.245359866" Nov 24 06:50:54 crc kubenswrapper[4799]: I1124 06:50:54.577122 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:54 crc kubenswrapper[4799]: I1124 06:50:54.577618 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:54 crc kubenswrapper[4799]: I1124 06:50:54.783906 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:54 crc kubenswrapper[4799]: I1124 06:50:54.783986 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:55 crc kubenswrapper[4799]: I1124 06:50:55.103794 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:50:55 crc kubenswrapper[4799]: I1124 06:50:55.104786 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:50:57 crc kubenswrapper[4799]: I1124 06:50:57.595216 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:57 crc kubenswrapper[4799]: I1124 06:50:57.595807 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:50:58 crc kubenswrapper[4799]: I1124 06:50:58.651939 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vzb6m" podUID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerName="registry-server" probeResult="failure" output=< Nov 24 06:50:58 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 06:50:58 crc kubenswrapper[4799]: > Nov 24 06:51:04 crc kubenswrapper[4799]: I1124 06:51:04.625818 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:51:04 crc kubenswrapper[4799]: I1124 06:51:04.842109 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:51:05 crc kubenswrapper[4799]: I1124 06:51:05.645629 4799 generic.go:334] "Generic (PLEG): container finished" podID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerID="0e108713267891d178bd7e0758c64aa03476be05bb71abb02c6cca6511fc13f7" exitCode=0 Nov 24 06:51:05 crc kubenswrapper[4799]: I1124 06:51:05.646102 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rqx5j" event={"ID":"e6461490-0f33-4730-83a1-7d01ac7f863b","Type":"ContainerDied","Data":"0e108713267891d178bd7e0758c64aa03476be05bb71abb02c6cca6511fc13f7"} Nov 24 06:51:05 crc kubenswrapper[4799]: I1124 06:51:05.650158 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csh4j" event={"ID":"3476fe21-9207-4c63-80e9-0c3e1bd7e19c","Type":"ContainerDied","Data":"a285a26d75fcf3fdc8a8b112f34c9a68abddb2227b2e97287cde497411cd06ea"} Nov 24 06:51:05 crc kubenswrapper[4799]: I1124 06:51:05.650158 4799 generic.go:334] "Generic (PLEG): container finished" podID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" containerID="a285a26d75fcf3fdc8a8b112f34c9a68abddb2227b2e97287cde497411cd06ea" exitCode=0 Nov 24 06:51:05 crc kubenswrapper[4799]: I1124 06:51:05.652103 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htg5n" event={"ID":"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a","Type":"ContainerStarted","Data":"0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad"} Nov 24 06:51:05 crc kubenswrapper[4799]: I1124 06:51:05.655343 4799 generic.go:334] "Generic (PLEG): container finished" podID="04184df2-64d2-4175-a1fd-a83670d7df6c" containerID="efc50a951e957903bf3d0ce90d2dd6298af45a8ce2f382425637db2f4875af2c" exitCode=0 Nov 24 06:51:05 crc kubenswrapper[4799]: I1124 06:51:05.655408 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvkkn" event={"ID":"04184df2-64d2-4175-a1fd-a83670d7df6c","Type":"ContainerDied","Data":"efc50a951e957903bf3d0ce90d2dd6298af45a8ce2f382425637db2f4875af2c"} Nov 24 06:51:05 crc kubenswrapper[4799]: I1124 06:51:05.861180 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sglfw"] Nov 24 06:51:05 crc kubenswrapper[4799]: I1124 06:51:05.861476 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sglfw" podUID="ea43b679-a4d2-4539-bcf3-130fbfe86656" containerName="registry-server" containerID="cri-o://53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5" gracePeriod=2 Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.297176 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.421383 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6fl2\" (UniqueName: \"kubernetes.io/projected/ea43b679-a4d2-4539-bcf3-130fbfe86656-kube-api-access-f6fl2\") pod \"ea43b679-a4d2-4539-bcf3-130fbfe86656\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.421526 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-catalog-content\") pod \"ea43b679-a4d2-4539-bcf3-130fbfe86656\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.421554 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-utilities\") pod \"ea43b679-a4d2-4539-bcf3-130fbfe86656\" (UID: \"ea43b679-a4d2-4539-bcf3-130fbfe86656\") " Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.425980 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-utilities" (OuterVolumeSpecName: "utilities") pod "ea43b679-a4d2-4539-bcf3-130fbfe86656" (UID: "ea43b679-a4d2-4539-bcf3-130fbfe86656"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.437035 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea43b679-a4d2-4539-bcf3-130fbfe86656-kube-api-access-f6fl2" (OuterVolumeSpecName: "kube-api-access-f6fl2") pod "ea43b679-a4d2-4539-bcf3-130fbfe86656" (UID: "ea43b679-a4d2-4539-bcf3-130fbfe86656"). InnerVolumeSpecName "kube-api-access-f6fl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.494906 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea43b679-a4d2-4539-bcf3-130fbfe86656" (UID: "ea43b679-a4d2-4539-bcf3-130fbfe86656"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.536972 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6fl2\" (UniqueName: \"kubernetes.io/projected/ea43b679-a4d2-4539-bcf3-130fbfe86656-kube-api-access-f6fl2\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.537022 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.537033 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea43b679-a4d2-4539-bcf3-130fbfe86656-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.623999 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gmz9"] Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.702755 4799 generic.go:334] "Generic (PLEG): container finished" podID="31145cc8-41cc-4870-83f8-bca24095bb0c" containerID="a1abba96b113431e8ebc0c2ecd30c91a78e1d517a69fe62cf50778e9e7d4f500" exitCode=0 Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.702943 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-288hz" event={"ID":"31145cc8-41cc-4870-83f8-bca24095bb0c","Type":"ContainerDied","Data":"a1abba96b113431e8ebc0c2ecd30c91a78e1d517a69fe62cf50778e9e7d4f500"} Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.727027 4799 generic.go:334] "Generic (PLEG): container finished" podID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" containerID="0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad" exitCode=0 Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.727110 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htg5n" event={"ID":"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a","Type":"ContainerDied","Data":"0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad"} Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.744590 4799 generic.go:334] "Generic (PLEG): container finished" podID="ea43b679-a4d2-4539-bcf3-130fbfe86656" containerID="53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5" exitCode=0 Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.744652 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sglfw" event={"ID":"ea43b679-a4d2-4539-bcf3-130fbfe86656","Type":"ContainerDied","Data":"53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5"} Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.744698 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sglfw" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.744711 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sglfw" event={"ID":"ea43b679-a4d2-4539-bcf3-130fbfe86656","Type":"ContainerDied","Data":"cd009638c8dde9be47a5f963ff3690b94816838f7c213fd8a036b3b68393b87b"} Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.744735 4799 scope.go:117] "RemoveContainer" containerID="53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.771902 4799 scope.go:117] "RemoveContainer" containerID="a8b369f081f419cd8e3afc97c9706d4679d30073243402601df5922ca1d450eb" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.793257 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sglfw"] Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.795758 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sglfw"] Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.813343 4799 scope.go:117] "RemoveContainer" containerID="f21d29de5814bec4cedc52e0f1441c6750e24186bb2ef8663700a14f765c3bf9" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.832514 4799 scope.go:117] "RemoveContainer" containerID="53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5" Nov 24 06:51:06 crc kubenswrapper[4799]: E1124 06:51:06.835979 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5\": container with ID starting with 53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5 not found: ID does not exist" containerID="53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.836040 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5"} err="failed to get container status \"53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5\": rpc error: code = NotFound desc = could not find container \"53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5\": container with ID starting with 53245238d444d32e1fc08dde25d3a202a129f32016298cd4dd3f5ba00fcc98a5 not found: ID does not exist" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.836113 4799 scope.go:117] "RemoveContainer" containerID="a8b369f081f419cd8e3afc97c9706d4679d30073243402601df5922ca1d450eb" Nov 24 06:51:06 crc kubenswrapper[4799]: E1124 06:51:06.839937 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8b369f081f419cd8e3afc97c9706d4679d30073243402601df5922ca1d450eb\": container with ID starting with a8b369f081f419cd8e3afc97c9706d4679d30073243402601df5922ca1d450eb not found: ID does not exist" containerID="a8b369f081f419cd8e3afc97c9706d4679d30073243402601df5922ca1d450eb" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.839973 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8b369f081f419cd8e3afc97c9706d4679d30073243402601df5922ca1d450eb"} err="failed to get container status \"a8b369f081f419cd8e3afc97c9706d4679d30073243402601df5922ca1d450eb\": rpc error: code = NotFound desc = could not find container \"a8b369f081f419cd8e3afc97c9706d4679d30073243402601df5922ca1d450eb\": container with ID starting with a8b369f081f419cd8e3afc97c9706d4679d30073243402601df5922ca1d450eb not found: ID does not exist" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.840001 4799 scope.go:117] "RemoveContainer" containerID="f21d29de5814bec4cedc52e0f1441c6750e24186bb2ef8663700a14f765c3bf9" Nov 24 06:51:06 crc kubenswrapper[4799]: E1124 06:51:06.841984 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f21d29de5814bec4cedc52e0f1441c6750e24186bb2ef8663700a14f765c3bf9\": container with ID starting with f21d29de5814bec4cedc52e0f1441c6750e24186bb2ef8663700a14f765c3bf9 not found: ID does not exist" containerID="f21d29de5814bec4cedc52e0f1441c6750e24186bb2ef8663700a14f765c3bf9" Nov 24 06:51:06 crc kubenswrapper[4799]: I1124 06:51:06.842017 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21d29de5814bec4cedc52e0f1441c6750e24186bb2ef8663700a14f765c3bf9"} err="failed to get container status \"f21d29de5814bec4cedc52e0f1441c6750e24186bb2ef8663700a14f765c3bf9\": rpc error: code = NotFound desc = could not find container \"f21d29de5814bec4cedc52e0f1441c6750e24186bb2ef8663700a14f765c3bf9\": container with ID starting with f21d29de5814bec4cedc52e0f1441c6750e24186bb2ef8663700a14f765c3bf9 not found: ID does not exist" Nov 24 06:51:07 crc kubenswrapper[4799]: I1124 06:51:07.642432 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea43b679-a4d2-4539-bcf3-130fbfe86656" path="/var/lib/kubelet/pods/ea43b679-a4d2-4539-bcf3-130fbfe86656/volumes" Nov 24 06:51:07 crc kubenswrapper[4799]: I1124 06:51:07.643552 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:51:07 crc kubenswrapper[4799]: I1124 06:51:07.702477 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:51:07 crc kubenswrapper[4799]: I1124 06:51:07.776892 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvkkn" event={"ID":"04184df2-64d2-4175-a1fd-a83670d7df6c","Type":"ContainerStarted","Data":"fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd"} Nov 24 06:51:07 crc kubenswrapper[4799]: I1124 06:51:07.782255 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rqx5j" event={"ID":"e6461490-0f33-4730-83a1-7d01ac7f863b","Type":"ContainerStarted","Data":"786576a1c16e129fb52d4aba744aa316ef23171c586945edf036b3ac0422156c"} Nov 24 06:51:07 crc kubenswrapper[4799]: I1124 06:51:07.784585 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csh4j" event={"ID":"3476fe21-9207-4c63-80e9-0c3e1bd7e19c","Type":"ContainerStarted","Data":"b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574"} Nov 24 06:51:07 crc kubenswrapper[4799]: I1124 06:51:07.797817 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pvkkn" podStartSLOduration=4.405538722 podStartE2EDuration="1m3.797794599s" podCreationTimestamp="2025-11-24 06:50:04 +0000 UTC" firstStartedPulling="2025-11-24 06:50:06.910262446 +0000 UTC m=+152.566244920" lastFinishedPulling="2025-11-24 06:51:06.302518313 +0000 UTC m=+211.958500797" observedRunningTime="2025-11-24 06:51:07.79533735 +0000 UTC m=+213.451319824" watchObservedRunningTime="2025-11-24 06:51:07.797794599 +0000 UTC m=+213.453777063" Nov 24 06:51:07 crc kubenswrapper[4799]: I1124 06:51:07.862586 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-csh4j" podStartSLOduration=4.579205253 podStartE2EDuration="1m3.86255791s" podCreationTimestamp="2025-11-24 06:50:04 +0000 UTC" firstStartedPulling="2025-11-24 06:50:06.898163907 +0000 UTC m=+152.554146381" lastFinishedPulling="2025-11-24 06:51:06.181516564 +0000 UTC m=+211.837499038" observedRunningTime="2025-11-24 06:51:07.8597783 +0000 UTC m=+213.515760774" watchObservedRunningTime="2025-11-24 06:51:07.86255791 +0000 UTC m=+213.518540384" Nov 24 06:51:07 crc kubenswrapper[4799]: I1124 06:51:07.863039 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rqx5j" podStartSLOduration=3.6949648489999998 podStartE2EDuration="1m0.863035005s" podCreationTimestamp="2025-11-24 06:50:07 +0000 UTC" firstStartedPulling="2025-11-24 06:50:09.03434615 +0000 UTC m=+154.690328624" lastFinishedPulling="2025-11-24 06:51:06.202416306 +0000 UTC m=+211.858398780" observedRunningTime="2025-11-24 06:51:07.83112426 +0000 UTC m=+213.487106754" watchObservedRunningTime="2025-11-24 06:51:07.863035005 +0000 UTC m=+213.519017479" Nov 24 06:51:07 crc kubenswrapper[4799]: I1124 06:51:07.993625 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:51:07 crc kubenswrapper[4799]: I1124 06:51:07.994145 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:51:08 crc kubenswrapper[4799]: I1124 06:51:08.792333 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-288hz" event={"ID":"31145cc8-41cc-4870-83f8-bca24095bb0c","Type":"ContainerStarted","Data":"8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a"} Nov 24 06:51:08 crc kubenswrapper[4799]: I1124 06:51:08.795702 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htg5n" event={"ID":"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a","Type":"ContainerStarted","Data":"8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf"} Nov 24 06:51:08 crc kubenswrapper[4799]: I1124 06:51:08.815887 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-288hz" podStartSLOduration=3.187892107 podStartE2EDuration="1m2.815861678s" podCreationTimestamp="2025-11-24 06:50:06 +0000 UTC" firstStartedPulling="2025-11-24 06:50:08.016941975 +0000 UTC m=+153.672924439" lastFinishedPulling="2025-11-24 06:51:07.644911536 +0000 UTC m=+213.300894010" observedRunningTime="2025-11-24 06:51:08.812640385 +0000 UTC m=+214.468622859" watchObservedRunningTime="2025-11-24 06:51:08.815861678 +0000 UTC m=+214.471844162" Nov 24 06:51:08 crc kubenswrapper[4799]: I1124 06:51:08.839803 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-htg5n" podStartSLOduration=3.030453093 podStartE2EDuration="1m2.839771887s" podCreationTimestamp="2025-11-24 06:50:06 +0000 UTC" firstStartedPulling="2025-11-24 06:50:07.989973644 +0000 UTC m=+153.645956118" lastFinishedPulling="2025-11-24 06:51:07.799292438 +0000 UTC m=+213.455274912" observedRunningTime="2025-11-24 06:51:08.8367629 +0000 UTC m=+214.492745374" watchObservedRunningTime="2025-11-24 06:51:08.839771887 +0000 UTC m=+214.495754351" Nov 24 06:51:09 crc kubenswrapper[4799]: I1124 06:51:09.028986 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rqx5j" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerName="registry-server" probeResult="failure" output=< Nov 24 06:51:09 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 06:51:09 crc kubenswrapper[4799]: > Nov 24 06:51:14 crc kubenswrapper[4799]: I1124 06:51:14.636657 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:51:14 crc kubenswrapper[4799]: I1124 06:51:14.638119 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:51:14 crc kubenswrapper[4799]: I1124 06:51:14.690707 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:51:14 crc kubenswrapper[4799]: I1124 06:51:14.882076 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:51:15 crc kubenswrapper[4799]: I1124 06:51:15.121165 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:51:15 crc kubenswrapper[4799]: I1124 06:51:15.121225 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:51:15 crc kubenswrapper[4799]: I1124 06:51:15.161220 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:51:15 crc kubenswrapper[4799]: I1124 06:51:15.892946 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:51:16 crc kubenswrapper[4799]: I1124 06:51:16.632726 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:51:16 crc kubenswrapper[4799]: I1124 06:51:16.632808 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:51:16 crc kubenswrapper[4799]: I1124 06:51:16.769220 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:51:16 crc kubenswrapper[4799]: I1124 06:51:16.889586 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:51:16 crc kubenswrapper[4799]: I1124 06:51:16.998881 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:51:16 crc kubenswrapper[4799]: I1124 06:51:16.999769 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:51:17 crc kubenswrapper[4799]: I1124 06:51:17.037376 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:51:17 crc kubenswrapper[4799]: I1124 06:51:17.274762 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pvkkn"] Nov 24 06:51:17 crc kubenswrapper[4799]: I1124 06:51:17.844740 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pvkkn" podUID="04184df2-64d2-4175-a1fd-a83670d7df6c" containerName="registry-server" containerID="cri-o://fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd" gracePeriod=2 Nov 24 06:51:17 crc kubenswrapper[4799]: I1124 06:51:17.906276 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:51:18 crc kubenswrapper[4799]: I1124 06:51:18.043908 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:51:18 crc kubenswrapper[4799]: I1124 06:51:18.114173 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.063929 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-htg5n"] Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.675628 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.741780 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6xks\" (UniqueName: \"kubernetes.io/projected/04184df2-64d2-4175-a1fd-a83670d7df6c-kube-api-access-n6xks\") pod \"04184df2-64d2-4175-a1fd-a83670d7df6c\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.742212 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-catalog-content\") pod \"04184df2-64d2-4175-a1fd-a83670d7df6c\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.742462 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-utilities\") pod \"04184df2-64d2-4175-a1fd-a83670d7df6c\" (UID: \"04184df2-64d2-4175-a1fd-a83670d7df6c\") " Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.743266 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-utilities" (OuterVolumeSpecName: "utilities") pod "04184df2-64d2-4175-a1fd-a83670d7df6c" (UID: "04184df2-64d2-4175-a1fd-a83670d7df6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.749111 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04184df2-64d2-4175-a1fd-a83670d7df6c-kube-api-access-n6xks" (OuterVolumeSpecName: "kube-api-access-n6xks") pod "04184df2-64d2-4175-a1fd-a83670d7df6c" (UID: "04184df2-64d2-4175-a1fd-a83670d7df6c"). InnerVolumeSpecName "kube-api-access-n6xks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.791096 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04184df2-64d2-4175-a1fd-a83670d7df6c" (UID: "04184df2-64d2-4175-a1fd-a83670d7df6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.844766 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.844840 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04184df2-64d2-4175-a1fd-a83670d7df6c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.844881 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6xks\" (UniqueName: \"kubernetes.io/projected/04184df2-64d2-4175-a1fd-a83670d7df6c-kube-api-access-n6xks\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.864904 4799 generic.go:334] "Generic (PLEG): container finished" podID="04184df2-64d2-4175-a1fd-a83670d7df6c" containerID="fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd" exitCode=0 Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.864991 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvkkn" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.865042 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvkkn" event={"ID":"04184df2-64d2-4175-a1fd-a83670d7df6c","Type":"ContainerDied","Data":"fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd"} Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.865082 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvkkn" event={"ID":"04184df2-64d2-4175-a1fd-a83670d7df6c","Type":"ContainerDied","Data":"48e0ecedb19e609fceb32bbc15575b37771947a3b082376dcc434c6a4dd0c8c2"} Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.865113 4799 scope.go:117] "RemoveContainer" containerID="fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.891576 4799 scope.go:117] "RemoveContainer" containerID="efc50a951e957903bf3d0ce90d2dd6298af45a8ce2f382425637db2f4875af2c" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.931501 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pvkkn"] Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.937081 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pvkkn"] Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.939251 4799 scope.go:117] "RemoveContainer" containerID="8539cae693510d551953b7ebacd79e759935d9c7a486993d561a55ad7e584a18" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.963271 4799 scope.go:117] "RemoveContainer" containerID="fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd" Nov 24 06:51:19 crc kubenswrapper[4799]: E1124 06:51:19.965020 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd\": container with ID starting with fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd not found: ID does not exist" containerID="fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.965059 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd"} err="failed to get container status \"fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd\": rpc error: code = NotFound desc = could not find container \"fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd\": container with ID starting with fcc9592bbbcc40544bc361ea5e858c2cfb5d643875ac2f26a697d7039ed523bd not found: ID does not exist" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.965093 4799 scope.go:117] "RemoveContainer" containerID="efc50a951e957903bf3d0ce90d2dd6298af45a8ce2f382425637db2f4875af2c" Nov 24 06:51:19 crc kubenswrapper[4799]: E1124 06:51:19.966191 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efc50a951e957903bf3d0ce90d2dd6298af45a8ce2f382425637db2f4875af2c\": container with ID starting with efc50a951e957903bf3d0ce90d2dd6298af45a8ce2f382425637db2f4875af2c not found: ID does not exist" containerID="efc50a951e957903bf3d0ce90d2dd6298af45a8ce2f382425637db2f4875af2c" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.966228 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efc50a951e957903bf3d0ce90d2dd6298af45a8ce2f382425637db2f4875af2c"} err="failed to get container status \"efc50a951e957903bf3d0ce90d2dd6298af45a8ce2f382425637db2f4875af2c\": rpc error: code = NotFound desc = could not find container \"efc50a951e957903bf3d0ce90d2dd6298af45a8ce2f382425637db2f4875af2c\": container with ID starting with efc50a951e957903bf3d0ce90d2dd6298af45a8ce2f382425637db2f4875af2c not found: ID does not exist" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.966250 4799 scope.go:117] "RemoveContainer" containerID="8539cae693510d551953b7ebacd79e759935d9c7a486993d561a55ad7e584a18" Nov 24 06:51:19 crc kubenswrapper[4799]: E1124 06:51:19.966681 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8539cae693510d551953b7ebacd79e759935d9c7a486993d561a55ad7e584a18\": container with ID starting with 8539cae693510d551953b7ebacd79e759935d9c7a486993d561a55ad7e584a18 not found: ID does not exist" containerID="8539cae693510d551953b7ebacd79e759935d9c7a486993d561a55ad7e584a18" Nov 24 06:51:19 crc kubenswrapper[4799]: I1124 06:51:19.966787 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8539cae693510d551953b7ebacd79e759935d9c7a486993d561a55ad7e584a18"} err="failed to get container status \"8539cae693510d551953b7ebacd79e759935d9c7a486993d561a55ad7e584a18\": rpc error: code = NotFound desc = could not find container \"8539cae693510d551953b7ebacd79e759935d9c7a486993d561a55ad7e584a18\": container with ID starting with 8539cae693510d551953b7ebacd79e759935d9c7a486993d561a55ad7e584a18 not found: ID does not exist" Nov 24 06:51:20 crc kubenswrapper[4799]: I1124 06:51:20.401152 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:51:20 crc kubenswrapper[4799]: I1124 06:51:20.401240 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:51:20 crc kubenswrapper[4799]: I1124 06:51:20.401301 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:51:20 crc kubenswrapper[4799]: I1124 06:51:20.402080 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 06:51:20 crc kubenswrapper[4799]: I1124 06:51:20.402146 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9" gracePeriod=600 Nov 24 06:51:20 crc kubenswrapper[4799]: I1124 06:51:20.879590 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9" exitCode=0 Nov 24 06:51:20 crc kubenswrapper[4799]: I1124 06:51:20.879734 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9"} Nov 24 06:51:20 crc kubenswrapper[4799]: I1124 06:51:20.880559 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"81921defcb796b9a98513b3ce531e00df43592241cc2ef5d7a1b9452de61e2c8"} Nov 24 06:51:20 crc kubenswrapper[4799]: I1124 06:51:20.881383 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-htg5n" podUID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" containerName="registry-server" containerID="cri-o://8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf" gracePeriod=2 Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.468695 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rqx5j"] Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.471411 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rqx5j" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerName="registry-server" containerID="cri-o://786576a1c16e129fb52d4aba744aa316ef23171c586945edf036b3ac0422156c" gracePeriod=2 Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.635149 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04184df2-64d2-4175-a1fd-a83670d7df6c" path="/var/lib/kubelet/pods/04184df2-64d2-4175-a1fd-a83670d7df6c/volumes" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.873438 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.889366 4799 generic.go:334] "Generic (PLEG): container finished" podID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" containerID="8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf" exitCode=0 Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.889474 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htg5n" event={"ID":"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a","Type":"ContainerDied","Data":"8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf"} Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.889519 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-htg5n" event={"ID":"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a","Type":"ContainerDied","Data":"bb003b979b928fd26af0774325783a3f319efd3f7dffcf748f33080a52cf8b25"} Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.889548 4799 scope.go:117] "RemoveContainer" containerID="8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.889690 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-htg5n" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.893011 4799 generic.go:334] "Generic (PLEG): container finished" podID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerID="786576a1c16e129fb52d4aba744aa316ef23171c586945edf036b3ac0422156c" exitCode=0 Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.893066 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rqx5j" event={"ID":"e6461490-0f33-4730-83a1-7d01ac7f863b","Type":"ContainerDied","Data":"786576a1c16e129fb52d4aba744aa316ef23171c586945edf036b3ac0422156c"} Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.909154 4799 scope.go:117] "RemoveContainer" containerID="0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.927274 4799 scope.go:117] "RemoveContainer" containerID="fda22a58e5759c18bb559846e28eb163132efbd27783e06e3600e9a0f8692be5" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.949213 4799 scope.go:117] "RemoveContainer" containerID="8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf" Nov 24 06:51:21 crc kubenswrapper[4799]: E1124 06:51:21.952181 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf\": container with ID starting with 8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf not found: ID does not exist" containerID="8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.952259 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf"} err="failed to get container status \"8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf\": rpc error: code = NotFound desc = could not find container \"8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf\": container with ID starting with 8a51818e2e1548f4e54cc21d51a76d505ad4f97b5327641407869fb752f12ebf not found: ID does not exist" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.952310 4799 scope.go:117] "RemoveContainer" containerID="0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad" Nov 24 06:51:21 crc kubenswrapper[4799]: E1124 06:51:21.952858 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad\": container with ID starting with 0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad not found: ID does not exist" containerID="0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.952893 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad"} err="failed to get container status \"0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad\": rpc error: code = NotFound desc = could not find container \"0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad\": container with ID starting with 0c7c4da1d47f7f5ff1da34e1eb08702b24671faba955f7506afb737837bbb0ad not found: ID does not exist" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.952912 4799 scope.go:117] "RemoveContainer" containerID="fda22a58e5759c18bb559846e28eb163132efbd27783e06e3600e9a0f8692be5" Nov 24 06:51:21 crc kubenswrapper[4799]: E1124 06:51:21.953272 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fda22a58e5759c18bb559846e28eb163132efbd27783e06e3600e9a0f8692be5\": container with ID starting with fda22a58e5759c18bb559846e28eb163132efbd27783e06e3600e9a0f8692be5 not found: ID does not exist" containerID="fda22a58e5759c18bb559846e28eb163132efbd27783e06e3600e9a0f8692be5" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.953302 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fda22a58e5759c18bb559846e28eb163132efbd27783e06e3600e9a0f8692be5"} err="failed to get container status \"fda22a58e5759c18bb559846e28eb163132efbd27783e06e3600e9a0f8692be5\": rpc error: code = NotFound desc = could not find container \"fda22a58e5759c18bb559846e28eb163132efbd27783e06e3600e9a0f8692be5\": container with ID starting with fda22a58e5759c18bb559846e28eb163132efbd27783e06e3600e9a0f8692be5 not found: ID does not exist" Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.994956 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-catalog-content\") pod \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.995061 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-utilities\") pod \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.995167 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8ffn\" (UniqueName: \"kubernetes.io/projected/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-kube-api-access-v8ffn\") pod \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\" (UID: \"e0cc7caf-7a38-4cd6-a809-098d38b8cf3a\") " Nov 24 06:51:21 crc kubenswrapper[4799]: I1124 06:51:21.996153 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-utilities" (OuterVolumeSpecName: "utilities") pod "e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" (UID: "e0cc7caf-7a38-4cd6-a809-098d38b8cf3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.009252 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-kube-api-access-v8ffn" (OuterVolumeSpecName: "kube-api-access-v8ffn") pod "e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" (UID: "e0cc7caf-7a38-4cd6-a809-098d38b8cf3a"). InnerVolumeSpecName "kube-api-access-v8ffn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.011639 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" (UID: "e0cc7caf-7a38-4cd6-a809-098d38b8cf3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.096655 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8ffn\" (UniqueName: \"kubernetes.io/projected/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-kube-api-access-v8ffn\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.096714 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.096728 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.220472 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-htg5n"] Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.224806 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-htg5n"] Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.334623 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.402216 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-utilities\") pod \"e6461490-0f33-4730-83a1-7d01ac7f863b\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.402286 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-catalog-content\") pod \"e6461490-0f33-4730-83a1-7d01ac7f863b\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.402644 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n994\" (UniqueName: \"kubernetes.io/projected/e6461490-0f33-4730-83a1-7d01ac7f863b-kube-api-access-4n994\") pod \"e6461490-0f33-4730-83a1-7d01ac7f863b\" (UID: \"e6461490-0f33-4730-83a1-7d01ac7f863b\") " Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.402972 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-utilities" (OuterVolumeSpecName: "utilities") pod "e6461490-0f33-4730-83a1-7d01ac7f863b" (UID: "e6461490-0f33-4730-83a1-7d01ac7f863b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.408408 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6461490-0f33-4730-83a1-7d01ac7f863b-kube-api-access-4n994" (OuterVolumeSpecName: "kube-api-access-4n994") pod "e6461490-0f33-4730-83a1-7d01ac7f863b" (UID: "e6461490-0f33-4730-83a1-7d01ac7f863b"). InnerVolumeSpecName "kube-api-access-4n994". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.503948 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n994\" (UniqueName: \"kubernetes.io/projected/e6461490-0f33-4730-83a1-7d01ac7f863b-kube-api-access-4n994\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.503992 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.512476 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6461490-0f33-4730-83a1-7d01ac7f863b" (UID: "e6461490-0f33-4730-83a1-7d01ac7f863b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.605487 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6461490-0f33-4730-83a1-7d01ac7f863b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.902926 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rqx5j" event={"ID":"e6461490-0f33-4730-83a1-7d01ac7f863b","Type":"ContainerDied","Data":"eff00acff6928b0481371d3e1ab6445debe7cfcb3f4fc865f7b86ff066e39ad2"} Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.903000 4799 scope.go:117] "RemoveContainer" containerID="786576a1c16e129fb52d4aba744aa316ef23171c586945edf036b3ac0422156c" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.903011 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rqx5j" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.923529 4799 scope.go:117] "RemoveContainer" containerID="0e108713267891d178bd7e0758c64aa03476be05bb71abb02c6cca6511fc13f7" Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.935979 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rqx5j"] Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.941284 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rqx5j"] Nov 24 06:51:22 crc kubenswrapper[4799]: I1124 06:51:22.949719 4799 scope.go:117] "RemoveContainer" containerID="e7c2a2fbcce322aff186b83de04e2832885951194c8be888842552b7d0b5a8d4" Nov 24 06:51:23 crc kubenswrapper[4799]: I1124 06:51:23.643422 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" path="/var/lib/kubelet/pods/e0cc7caf-7a38-4cd6-a809-098d38b8cf3a/volumes" Nov 24 06:51:23 crc kubenswrapper[4799]: I1124 06:51:23.644961 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" path="/var/lib/kubelet/pods/e6461490-0f33-4730-83a1-7d01ac7f863b/volumes" Nov 24 06:51:31 crc kubenswrapper[4799]: I1124 06:51:31.657780 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" podUID="7dc41818-1f5e-4472-a726-f1322a1f462b" containerName="oauth-openshift" containerID="cri-o://fb4f1c9a2efff87d19c91ec02c301953d04f734c9102e857df83174952a28059" gracePeriod=15 Nov 24 06:51:31 crc kubenswrapper[4799]: I1124 06:51:31.984216 4799 generic.go:334] "Generic (PLEG): container finished" podID="7dc41818-1f5e-4472-a726-f1322a1f462b" containerID="fb4f1c9a2efff87d19c91ec02c301953d04f734c9102e857df83174952a28059" exitCode=0 Nov 24 06:51:31 crc kubenswrapper[4799]: I1124 06:51:31.984504 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" event={"ID":"7dc41818-1f5e-4472-a726-f1322a1f462b","Type":"ContainerDied","Data":"fb4f1c9a2efff87d19c91ec02c301953d04f734c9102e857df83174952a28059"} Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.099835 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167510 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-dir\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167593 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-cliconfig\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167642 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167660 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-serving-cert\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167727 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-idp-0-file-data\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167753 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-session\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167779 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-ocp-branding-template\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167809 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6bgh\" (UniqueName: \"kubernetes.io/projected/7dc41818-1f5e-4472-a726-f1322a1f462b-kube-api-access-l6bgh\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167835 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-provider-selection\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167893 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-policies\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167928 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-error\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167965 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-service-ca\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.167984 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-trusted-ca-bundle\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.168004 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-login\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.168036 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-router-certs\") pod \"7dc41818-1f5e-4472-a726-f1322a1f462b\" (UID: \"7dc41818-1f5e-4472-a726-f1322a1f462b\") " Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.168288 4799 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.168592 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.169559 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.169637 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.169676 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.175051 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.175326 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.175618 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dc41818-1f5e-4472-a726-f1322a1f462b-kube-api-access-l6bgh" (OuterVolumeSpecName: "kube-api-access-l6bgh") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "kube-api-access-l6bgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.176033 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.176455 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.179137 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.179588 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.180288 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.183406 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7dc41818-1f5e-4472-a726-f1322a1f462b" (UID: "7dc41818-1f5e-4472-a726-f1322a1f462b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269717 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269768 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269783 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269795 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269812 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269833 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6bgh\" (UniqueName: \"kubernetes.io/projected/7dc41818-1f5e-4472-a726-f1322a1f462b-kube-api-access-l6bgh\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269868 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269886 4799 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269897 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269912 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269926 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269942 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.269954 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7dc41818-1f5e-4472-a726-f1322a1f462b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.995447 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" event={"ID":"7dc41818-1f5e-4472-a726-f1322a1f462b","Type":"ContainerDied","Data":"6b1e2cd271677e8f1dc88e56e97fa4b0dca01f62a727bb324436b5810a142a94"} Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.996056 4799 scope.go:117] "RemoveContainer" containerID="fb4f1c9a2efff87d19c91ec02c301953d04f734c9102e857df83174952a28059" Nov 24 06:51:32 crc kubenswrapper[4799]: I1124 06:51:32.995582 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gmz9" Nov 24 06:51:33 crc kubenswrapper[4799]: I1124 06:51:33.051782 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gmz9"] Nov 24 06:51:33 crc kubenswrapper[4799]: I1124 06:51:33.057340 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gmz9"] Nov 24 06:51:33 crc kubenswrapper[4799]: I1124 06:51:33.641301 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dc41818-1f5e-4472-a726-f1322a1f462b" path="/var/lib/kubelet/pods/7dc41818-1f5e-4472-a726-f1322a1f462b/volumes" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013234 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-745f6bf96d-dd7q8"] Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013527 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04184df2-64d2-4175-a1fd-a83670d7df6c" containerName="extract-content" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013547 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="04184df2-64d2-4175-a1fd-a83670d7df6c" containerName="extract-content" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013563 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e116addf-dac0-42aa-9c0e-82383b214399" containerName="collect-profiles" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013573 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e116addf-dac0-42aa-9c0e-82383b214399" containerName="collect-profiles" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013584 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013592 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013604 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" containerName="extract-content" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013611 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" containerName="extract-content" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013623 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04184df2-64d2-4175-a1fd-a83670d7df6c" containerName="extract-utilities" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013631 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="04184df2-64d2-4175-a1fd-a83670d7df6c" containerName="extract-utilities" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013641 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed153767-6876-4cb0-9344-cfc0133c9ff8" containerName="pruner" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013647 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed153767-6876-4cb0-9344-cfc0133c9ff8" containerName="pruner" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013655 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerName="extract-utilities" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013664 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerName="extract-utilities" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013673 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea43b679-a4d2-4539-bcf3-130fbfe86656" containerName="extract-utilities" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013679 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea43b679-a4d2-4539-bcf3-130fbfe86656" containerName="extract-utilities" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013688 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" containerName="extract-utilities" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013694 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" containerName="extract-utilities" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013701 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea43b679-a4d2-4539-bcf3-130fbfe86656" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013707 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea43b679-a4d2-4539-bcf3-130fbfe86656" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013713 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013719 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013725 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc41818-1f5e-4472-a726-f1322a1f462b" containerName="oauth-openshift" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013731 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc41818-1f5e-4472-a726-f1322a1f462b" containerName="oauth-openshift" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013745 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerName="extract-content" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013753 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerName="extract-content" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013766 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04184df2-64d2-4175-a1fd-a83670d7df6c" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013779 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="04184df2-64d2-4175-a1fd-a83670d7df6c" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013788 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63282372-2d7d-40fe-b64a-adea4d6931e3" containerName="pruner" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013795 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="63282372-2d7d-40fe-b64a-adea4d6931e3" containerName="pruner" Nov 24 06:51:34 crc kubenswrapper[4799]: E1124 06:51:34.013804 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea43b679-a4d2-4539-bcf3-130fbfe86656" containerName="extract-content" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013810 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea43b679-a4d2-4539-bcf3-130fbfe86656" containerName="extract-content" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.013981 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e116addf-dac0-42aa-9c0e-82383b214399" containerName="collect-profiles" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.014004 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed153767-6876-4cb0-9344-cfc0133c9ff8" containerName="pruner" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.014017 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dc41818-1f5e-4472-a726-f1322a1f462b" containerName="oauth-openshift" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.014026 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6461490-0f33-4730-83a1-7d01ac7f863b" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.014038 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0cc7caf-7a38-4cd6-a809-098d38b8cf3a" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.014049 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea43b679-a4d2-4539-bcf3-130fbfe86656" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.014063 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="04184df2-64d2-4175-a1fd-a83670d7df6c" containerName="registry-server" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.014076 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="63282372-2d7d-40fe-b64a-adea4d6931e3" containerName="pruner" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.014544 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.018975 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.019612 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.020052 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.020366 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.020647 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.020782 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.020798 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.020882 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.021121 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.021221 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.021330 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.021959 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.040701 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.040787 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.045395 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-745f6bf96d-dd7q8"] Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.048685 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099250 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099357 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5858ef4f-3d77-4c4e-9e38-812111c93d49-audit-dir\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099393 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099469 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-audit-policies\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099518 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-cliconfig\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099543 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-service-ca\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099605 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-router-certs\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099631 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099654 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rdzk\" (UniqueName: \"kubernetes.io/projected/5858ef4f-3d77-4c4e-9e38-812111c93d49-kube-api-access-7rdzk\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099683 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-session\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099765 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-template-login\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.099937 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-template-error\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.100046 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-serving-cert\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.100074 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201384 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-template-login\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201477 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-template-error\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201536 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-serving-cert\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201579 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201656 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201700 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5858ef4f-3d77-4c4e-9e38-812111c93d49-audit-dir\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201737 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-audit-policies\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201771 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201808 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-cliconfig\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201875 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-service-ca\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201923 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-router-certs\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.201961 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.202008 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rdzk\" (UniqueName: \"kubernetes.io/projected/5858ef4f-3d77-4c4e-9e38-812111c93d49-kube-api-access-7rdzk\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.202102 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-session\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.203328 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-service-ca\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.203455 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5858ef4f-3d77-4c4e-9e38-812111c93d49-audit-dir\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.203611 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-audit-policies\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.203886 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-cliconfig\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.204184 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.209642 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-template-error\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.209960 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.210674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-serving-cert\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.211422 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-template-login\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.211908 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-router-certs\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.212129 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.214771 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-system-session\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.216579 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5858ef4f-3d77-4c4e-9e38-812111c93d49-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.232750 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rdzk\" (UniqueName: \"kubernetes.io/projected/5858ef4f-3d77-4c4e-9e38-812111c93d49-kube-api-access-7rdzk\") pod \"oauth-openshift-745f6bf96d-dd7q8\" (UID: \"5858ef4f-3d77-4c4e-9e38-812111c93d49\") " pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.357751 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:34 crc kubenswrapper[4799]: I1124 06:51:34.734509 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-745f6bf96d-dd7q8"] Nov 24 06:51:35 crc kubenswrapper[4799]: I1124 06:51:35.017683 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" event={"ID":"5858ef4f-3d77-4c4e-9e38-812111c93d49","Type":"ContainerStarted","Data":"fc79a33fc080391a02172960c156bcfa950bc8114e35b56d30d41620bf157a00"} Nov 24 06:51:36 crc kubenswrapper[4799]: I1124 06:51:36.029642 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" event={"ID":"5858ef4f-3d77-4c4e-9e38-812111c93d49","Type":"ContainerStarted","Data":"62d4331d43ea3ae5ed2fc8f4a05336e7f37d086e3c5cf5c3323d7c6ed6aba1c1"} Nov 24 06:51:36 crc kubenswrapper[4799]: I1124 06:51:36.030028 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:36 crc kubenswrapper[4799]: I1124 06:51:36.040005 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" Nov 24 06:51:36 crc kubenswrapper[4799]: I1124 06:51:36.070956 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-745f6bf96d-dd7q8" podStartSLOduration=30.070927608 podStartE2EDuration="30.070927608s" podCreationTimestamp="2025-11-24 06:51:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:51:36.069233894 +0000 UTC m=+241.725216378" watchObservedRunningTime="2025-11-24 06:51:36.070927608 +0000 UTC m=+241.726910092" Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.531727 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-csh4j"] Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.533128 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-csh4j" podUID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" containerName="registry-server" containerID="cri-o://b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574" gracePeriod=30 Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.542553 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r2hbg"] Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.546104 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r2hbg" podUID="787fe453-1786-449c-bf60-87153ce058f9" containerName="registry-server" containerID="cri-o://51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a" gracePeriod=30 Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.557764 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcvtw"] Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.558103 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" podUID="3b02c876-c14b-4422-beaf-9ea3ab910f90" containerName="marketplace-operator" containerID="cri-o://b4c7b3100a3fc2990475a72d5d9965c9288ff6e1f0c3c62513f32bcf494ab724" gracePeriod=30 Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.574539 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-288hz"] Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.574942 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-288hz" podUID="31145cc8-41cc-4870-83f8-bca24095bb0c" containerName="registry-server" containerID="cri-o://8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a" gracePeriod=30 Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.579555 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vzb6m"] Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.583682 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vzb6m" podUID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerName="registry-server" containerID="cri-o://ecf3144ce85fdf6e45be5912761a06823a3a942e244d8b1dd07d2dff1b728aff" gracePeriod=30 Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.610346 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2hdxh"] Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.611141 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.615583 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2hdxh"] Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.697692 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c598ccb-346a-45eb-ad78-ad538667a705-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2hdxh\" (UID: \"3c598ccb-346a-45eb-ad78-ad538667a705\") " pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.697789 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nkkc\" (UniqueName: \"kubernetes.io/projected/3c598ccb-346a-45eb-ad78-ad538667a705-kube-api-access-2nkkc\") pod \"marketplace-operator-79b997595-2hdxh\" (UID: \"3c598ccb-346a-45eb-ad78-ad538667a705\") " pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.697829 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3c598ccb-346a-45eb-ad78-ad538667a705-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2hdxh\" (UID: \"3c598ccb-346a-45eb-ad78-ad538667a705\") " pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.799578 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nkkc\" (UniqueName: \"kubernetes.io/projected/3c598ccb-346a-45eb-ad78-ad538667a705-kube-api-access-2nkkc\") pod \"marketplace-operator-79b997595-2hdxh\" (UID: \"3c598ccb-346a-45eb-ad78-ad538667a705\") " pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.799658 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3c598ccb-346a-45eb-ad78-ad538667a705-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2hdxh\" (UID: \"3c598ccb-346a-45eb-ad78-ad538667a705\") " pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.799723 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c598ccb-346a-45eb-ad78-ad538667a705-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2hdxh\" (UID: \"3c598ccb-346a-45eb-ad78-ad538667a705\") " pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.801677 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c598ccb-346a-45eb-ad78-ad538667a705-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2hdxh\" (UID: \"3c598ccb-346a-45eb-ad78-ad538667a705\") " pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.807705 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3c598ccb-346a-45eb-ad78-ad538667a705-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2hdxh\" (UID: \"3c598ccb-346a-45eb-ad78-ad538667a705\") " pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:48 crc kubenswrapper[4799]: I1124 06:51:48.819096 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nkkc\" (UniqueName: \"kubernetes.io/projected/3c598ccb-346a-45eb-ad78-ad538667a705-kube-api-access-2nkkc\") pod \"marketplace-operator-79b997595-2hdxh\" (UID: \"3c598ccb-346a-45eb-ad78-ad538667a705\") " pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.091059 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.094741 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.101469 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.109353 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.135606 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2hbg" event={"ID":"787fe453-1786-449c-bf60-87153ce058f9","Type":"ContainerDied","Data":"51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a"} Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.135689 4799 scope.go:117] "RemoveContainer" containerID="51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.135641 4799 generic.go:334] "Generic (PLEG): container finished" podID="787fe453-1786-449c-bf60-87153ce058f9" containerID="51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a" exitCode=0 Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.135794 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2hbg" event={"ID":"787fe453-1786-449c-bf60-87153ce058f9","Type":"ContainerDied","Data":"8b1f0c47b9a118f38140e805eca6a66aef24bfe127cf597c3e144878bad78472"} Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.136630 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2hbg" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.154680 4799 generic.go:334] "Generic (PLEG): container finished" podID="31145cc8-41cc-4870-83f8-bca24095bb0c" containerID="8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a" exitCode=0 Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.154784 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-288hz" event={"ID":"31145cc8-41cc-4870-83f8-bca24095bb0c","Type":"ContainerDied","Data":"8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a"} Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.154797 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-288hz" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.154821 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-288hz" event={"ID":"31145cc8-41cc-4870-83f8-bca24095bb0c","Type":"ContainerDied","Data":"9e233586edb956dfb147e59a6bb27b1fed603bca262d06844378f9f3bb6376e2"} Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.156822 4799 generic.go:334] "Generic (PLEG): container finished" podID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" containerID="b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574" exitCode=0 Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.156873 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csh4j" event={"ID":"3476fe21-9207-4c63-80e9-0c3e1bd7e19c","Type":"ContainerDied","Data":"b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574"} Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.156892 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csh4j" event={"ID":"3476fe21-9207-4c63-80e9-0c3e1bd7e19c","Type":"ContainerDied","Data":"74f6e98ea94738701935e11fa028e308b4ec56cbbb363cba2159eda110dc0d1e"} Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.156962 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csh4j" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.159469 4799 generic.go:334] "Generic (PLEG): container finished" podID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerID="ecf3144ce85fdf6e45be5912761a06823a3a942e244d8b1dd07d2dff1b728aff" exitCode=0 Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.159559 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzb6m" event={"ID":"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740","Type":"ContainerDied","Data":"ecf3144ce85fdf6e45be5912761a06823a3a942e244d8b1dd07d2dff1b728aff"} Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.165125 4799 scope.go:117] "RemoveContainer" containerID="1914af0f9a86a50d913a5f0c25653b449863cc18ee45f1b368cccb7044a7a06e" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.178180 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.181839 4799 generic.go:334] "Generic (PLEG): container finished" podID="3b02c876-c14b-4422-beaf-9ea3ab910f90" containerID="b4c7b3100a3fc2990475a72d5d9965c9288ff6e1f0c3c62513f32bcf494ab724" exitCode=0 Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.181916 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" event={"ID":"3b02c876-c14b-4422-beaf-9ea3ab910f90","Type":"ContainerDied","Data":"b4c7b3100a3fc2990475a72d5d9965c9288ff6e1f0c3c62513f32bcf494ab724"} Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.193796 4799 scope.go:117] "RemoveContainer" containerID="f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.213871 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-catalog-content\") pod \"31145cc8-41cc-4870-83f8-bca24095bb0c\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.214359 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvx26\" (UniqueName: \"kubernetes.io/projected/31145cc8-41cc-4870-83f8-bca24095bb0c-kube-api-access-fvx26\") pod \"31145cc8-41cc-4870-83f8-bca24095bb0c\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.214554 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-utilities\") pod \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.214953 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrk6v\" (UniqueName: \"kubernetes.io/projected/787fe453-1786-449c-bf60-87153ce058f9-kube-api-access-rrk6v\") pod \"787fe453-1786-449c-bf60-87153ce058f9\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.215102 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-utilities\") pod \"787fe453-1786-449c-bf60-87153ce058f9\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.215218 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-catalog-content\") pod \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.215381 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmdxc\" (UniqueName: \"kubernetes.io/projected/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-kube-api-access-nmdxc\") pod \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\" (UID: \"3476fe21-9207-4c63-80e9-0c3e1bd7e19c\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.215541 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-utilities\") pod \"31145cc8-41cc-4870-83f8-bca24095bb0c\" (UID: \"31145cc8-41cc-4870-83f8-bca24095bb0c\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.215679 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-catalog-content\") pod \"787fe453-1786-449c-bf60-87153ce058f9\" (UID: \"787fe453-1786-449c-bf60-87153ce058f9\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.216683 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-utilities" (OuterVolumeSpecName: "utilities") pod "3476fe21-9207-4c63-80e9-0c3e1bd7e19c" (UID: "3476fe21-9207-4c63-80e9-0c3e1bd7e19c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.216969 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-utilities" (OuterVolumeSpecName: "utilities") pod "787fe453-1786-449c-bf60-87153ce058f9" (UID: "787fe453-1786-449c-bf60-87153ce058f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.217590 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-utilities" (OuterVolumeSpecName: "utilities") pod "31145cc8-41cc-4870-83f8-bca24095bb0c" (UID: "31145cc8-41cc-4870-83f8-bca24095bb0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.223179 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/787fe453-1786-449c-bf60-87153ce058f9-kube-api-access-rrk6v" (OuterVolumeSpecName: "kube-api-access-rrk6v") pod "787fe453-1786-449c-bf60-87153ce058f9" (UID: "787fe453-1786-449c-bf60-87153ce058f9"). InnerVolumeSpecName "kube-api-access-rrk6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.223252 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-kube-api-access-nmdxc" (OuterVolumeSpecName: "kube-api-access-nmdxc") pod "3476fe21-9207-4c63-80e9-0c3e1bd7e19c" (UID: "3476fe21-9207-4c63-80e9-0c3e1bd7e19c"). InnerVolumeSpecName "kube-api-access-nmdxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.226610 4799 scope.go:117] "RemoveContainer" containerID="51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a" Nov 24 06:51:49 crc kubenswrapper[4799]: E1124 06:51:49.229960 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a\": container with ID starting with 51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a not found: ID does not exist" containerID="51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.230013 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a"} err="failed to get container status \"51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a\": rpc error: code = NotFound desc = could not find container \"51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a\": container with ID starting with 51ba8410abe3ce2e4258103924637717eea4db9c6611b9505aa2d5cfa124ae6a not found: ID does not exist" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.230073 4799 scope.go:117] "RemoveContainer" containerID="1914af0f9a86a50d913a5f0c25653b449863cc18ee45f1b368cccb7044a7a06e" Nov 24 06:51:49 crc kubenswrapper[4799]: E1124 06:51:49.230379 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1914af0f9a86a50d913a5f0c25653b449863cc18ee45f1b368cccb7044a7a06e\": container with ID starting with 1914af0f9a86a50d913a5f0c25653b449863cc18ee45f1b368cccb7044a7a06e not found: ID does not exist" containerID="1914af0f9a86a50d913a5f0c25653b449863cc18ee45f1b368cccb7044a7a06e" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.230400 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1914af0f9a86a50d913a5f0c25653b449863cc18ee45f1b368cccb7044a7a06e"} err="failed to get container status \"1914af0f9a86a50d913a5f0c25653b449863cc18ee45f1b368cccb7044a7a06e\": rpc error: code = NotFound desc = could not find container \"1914af0f9a86a50d913a5f0c25653b449863cc18ee45f1b368cccb7044a7a06e\": container with ID starting with 1914af0f9a86a50d913a5f0c25653b449863cc18ee45f1b368cccb7044a7a06e not found: ID does not exist" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.230416 4799 scope.go:117] "RemoveContainer" containerID="f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a" Nov 24 06:51:49 crc kubenswrapper[4799]: E1124 06:51:49.230675 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a\": container with ID starting with f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a not found: ID does not exist" containerID="f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.230704 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a"} err="failed to get container status \"f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a\": rpc error: code = NotFound desc = could not find container \"f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a\": container with ID starting with f995054579cd4f11d68970b6f3f1e37bd90e2a0cdf9347b86cb97a7e42b3a60a not found: ID does not exist" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.230725 4799 scope.go:117] "RemoveContainer" containerID="8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.233083 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.233106 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrk6v\" (UniqueName: \"kubernetes.io/projected/787fe453-1786-449c-bf60-87153ce058f9-kube-api-access-rrk6v\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.233118 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.233130 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmdxc\" (UniqueName: \"kubernetes.io/projected/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-kube-api-access-nmdxc\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.233142 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.240061 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31145cc8-41cc-4870-83f8-bca24095bb0c-kube-api-access-fvx26" (OuterVolumeSpecName: "kube-api-access-fvx26") pod "31145cc8-41cc-4870-83f8-bca24095bb0c" (UID: "31145cc8-41cc-4870-83f8-bca24095bb0c"). InnerVolumeSpecName "kube-api-access-fvx26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.247567 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31145cc8-41cc-4870-83f8-bca24095bb0c" (UID: "31145cc8-41cc-4870-83f8-bca24095bb0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.256544 4799 scope.go:117] "RemoveContainer" containerID="a1abba96b113431e8ebc0c2ecd30c91a78e1d517a69fe62cf50778e9e7d4f500" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.286498 4799 scope.go:117] "RemoveContainer" containerID="a85acdb86cfd13022e9cd801757268441a2382f918ccca53f464209c2696ed81" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.307343 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "787fe453-1786-449c-bf60-87153ce058f9" (UID: "787fe453-1786-449c-bf60-87153ce058f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.308276 4799 scope.go:117] "RemoveContainer" containerID="8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a" Nov 24 06:51:49 crc kubenswrapper[4799]: E1124 06:51:49.309095 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a\": container with ID starting with 8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a not found: ID does not exist" containerID="8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.309135 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a"} err="failed to get container status \"8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a\": rpc error: code = NotFound desc = could not find container \"8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a\": container with ID starting with 8711fba08058467c7ebe0c6928c8f1cc9f2ddc83f6ac9f50be06912769f73d8a not found: ID does not exist" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.309163 4799 scope.go:117] "RemoveContainer" containerID="a1abba96b113431e8ebc0c2ecd30c91a78e1d517a69fe62cf50778e9e7d4f500" Nov 24 06:51:49 crc kubenswrapper[4799]: E1124 06:51:49.309527 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1abba96b113431e8ebc0c2ecd30c91a78e1d517a69fe62cf50778e9e7d4f500\": container with ID starting with a1abba96b113431e8ebc0c2ecd30c91a78e1d517a69fe62cf50778e9e7d4f500 not found: ID does not exist" containerID="a1abba96b113431e8ebc0c2ecd30c91a78e1d517a69fe62cf50778e9e7d4f500" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.309585 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1abba96b113431e8ebc0c2ecd30c91a78e1d517a69fe62cf50778e9e7d4f500"} err="failed to get container status \"a1abba96b113431e8ebc0c2ecd30c91a78e1d517a69fe62cf50778e9e7d4f500\": rpc error: code = NotFound desc = could not find container \"a1abba96b113431e8ebc0c2ecd30c91a78e1d517a69fe62cf50778e9e7d4f500\": container with ID starting with a1abba96b113431e8ebc0c2ecd30c91a78e1d517a69fe62cf50778e9e7d4f500 not found: ID does not exist" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.309622 4799 scope.go:117] "RemoveContainer" containerID="a85acdb86cfd13022e9cd801757268441a2382f918ccca53f464209c2696ed81" Nov 24 06:51:49 crc kubenswrapper[4799]: E1124 06:51:49.310143 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a85acdb86cfd13022e9cd801757268441a2382f918ccca53f464209c2696ed81\": container with ID starting with a85acdb86cfd13022e9cd801757268441a2382f918ccca53f464209c2696ed81 not found: ID does not exist" containerID="a85acdb86cfd13022e9cd801757268441a2382f918ccca53f464209c2696ed81" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.310265 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a85acdb86cfd13022e9cd801757268441a2382f918ccca53f464209c2696ed81"} err="failed to get container status \"a85acdb86cfd13022e9cd801757268441a2382f918ccca53f464209c2696ed81\": rpc error: code = NotFound desc = could not find container \"a85acdb86cfd13022e9cd801757268441a2382f918ccca53f464209c2696ed81\": container with ID starting with a85acdb86cfd13022e9cd801757268441a2382f918ccca53f464209c2696ed81 not found: ID does not exist" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.310655 4799 scope.go:117] "RemoveContainer" containerID="b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.324175 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3476fe21-9207-4c63-80e9-0c3e1bd7e19c" (UID: "3476fe21-9207-4c63-80e9-0c3e1bd7e19c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.330331 4799 scope.go:117] "RemoveContainer" containerID="a285a26d75fcf3fdc8a8b112f34c9a68abddb2227b2e97287cde497411cd06ea" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.333942 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ftkp\" (UniqueName: \"kubernetes.io/projected/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-kube-api-access-9ftkp\") pod \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.333991 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-catalog-content\") pod \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.334031 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-utilities\") pod \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\" (UID: \"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.334186 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/787fe453-1786-449c-bf60-87153ce058f9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.334204 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31145cc8-41cc-4870-83f8-bca24095bb0c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.334213 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvx26\" (UniqueName: \"kubernetes.io/projected/31145cc8-41cc-4870-83f8-bca24095bb0c-kube-api-access-fvx26\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.334225 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3476fe21-9207-4c63-80e9-0c3e1bd7e19c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.334794 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-utilities" (OuterVolumeSpecName: "utilities") pod "7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" (UID: "7dab8ddb-4cb5-4349-a4e5-22b50fbf0740"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.338370 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-kube-api-access-9ftkp" (OuterVolumeSpecName: "kube-api-access-9ftkp") pod "7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" (UID: "7dab8ddb-4cb5-4349-a4e5-22b50fbf0740"). InnerVolumeSpecName "kube-api-access-9ftkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.347711 4799 scope.go:117] "RemoveContainer" containerID="73b29289760a17b5c99878355b59e3c29d1c7de4850786852fb28091cb8875ce" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.375121 4799 scope.go:117] "RemoveContainer" containerID="b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574" Nov 24 06:51:49 crc kubenswrapper[4799]: E1124 06:51:49.376019 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574\": container with ID starting with b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574 not found: ID does not exist" containerID="b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.376057 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574"} err="failed to get container status \"b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574\": rpc error: code = NotFound desc = could not find container \"b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574\": container with ID starting with b407728f697ef9119dc134813dc2d631e090de8f910b3ada775f0df8d22f4574 not found: ID does not exist" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.376086 4799 scope.go:117] "RemoveContainer" containerID="a285a26d75fcf3fdc8a8b112f34c9a68abddb2227b2e97287cde497411cd06ea" Nov 24 06:51:49 crc kubenswrapper[4799]: E1124 06:51:49.376642 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a285a26d75fcf3fdc8a8b112f34c9a68abddb2227b2e97287cde497411cd06ea\": container with ID starting with a285a26d75fcf3fdc8a8b112f34c9a68abddb2227b2e97287cde497411cd06ea not found: ID does not exist" containerID="a285a26d75fcf3fdc8a8b112f34c9a68abddb2227b2e97287cde497411cd06ea" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.376702 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a285a26d75fcf3fdc8a8b112f34c9a68abddb2227b2e97287cde497411cd06ea"} err="failed to get container status \"a285a26d75fcf3fdc8a8b112f34c9a68abddb2227b2e97287cde497411cd06ea\": rpc error: code = NotFound desc = could not find container \"a285a26d75fcf3fdc8a8b112f34c9a68abddb2227b2e97287cde497411cd06ea\": container with ID starting with a285a26d75fcf3fdc8a8b112f34c9a68abddb2227b2e97287cde497411cd06ea not found: ID does not exist" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.376752 4799 scope.go:117] "RemoveContainer" containerID="73b29289760a17b5c99878355b59e3c29d1c7de4850786852fb28091cb8875ce" Nov 24 06:51:49 crc kubenswrapper[4799]: E1124 06:51:49.377316 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73b29289760a17b5c99878355b59e3c29d1c7de4850786852fb28091cb8875ce\": container with ID starting with 73b29289760a17b5c99878355b59e3c29d1c7de4850786852fb28091cb8875ce not found: ID does not exist" containerID="73b29289760a17b5c99878355b59e3c29d1c7de4850786852fb28091cb8875ce" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.377340 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b29289760a17b5c99878355b59e3c29d1c7de4850786852fb28091cb8875ce"} err="failed to get container status \"73b29289760a17b5c99878355b59e3c29d1c7de4850786852fb28091cb8875ce\": rpc error: code = NotFound desc = could not find container \"73b29289760a17b5c99878355b59e3c29d1c7de4850786852fb28091cb8875ce\": container with ID starting with 73b29289760a17b5c99878355b59e3c29d1c7de4850786852fb28091cb8875ce not found: ID does not exist" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.435062 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ftkp\" (UniqueName: \"kubernetes.io/projected/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-kube-api-access-9ftkp\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.435103 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.466781 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.471551 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" (UID: "7dab8ddb-4cb5-4349-a4e5-22b50fbf0740"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.475670 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r2hbg"] Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.478415 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r2hbg"] Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.511163 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-288hz"] Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.520548 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-288hz"] Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.531692 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-csh4j"] Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.535444 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-csh4j"] Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.536186 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.557589 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2hdxh"] Nov 24 06:51:49 crc kubenswrapper[4799]: W1124 06:51:49.563692 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c598ccb_346a_45eb_ad78_ad538667a705.slice/crio-1e6f63c2c0e9d2903fdcb97a8001e8522446d1e05420d9c36892acb6be5a7a67 WatchSource:0}: Error finding container 1e6f63c2c0e9d2903fdcb97a8001e8522446d1e05420d9c36892acb6be5a7a67: Status 404 returned error can't find the container with id 1e6f63c2c0e9d2903fdcb97a8001e8522446d1e05420d9c36892acb6be5a7a67 Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.640836 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-trusted-ca\") pod \"3b02c876-c14b-4422-beaf-9ea3ab910f90\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.641530 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzrrh\" (UniqueName: \"kubernetes.io/projected/3b02c876-c14b-4422-beaf-9ea3ab910f90-kube-api-access-xzrrh\") pod \"3b02c876-c14b-4422-beaf-9ea3ab910f90\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.641585 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-operator-metrics\") pod \"3b02c876-c14b-4422-beaf-9ea3ab910f90\" (UID: \"3b02c876-c14b-4422-beaf-9ea3ab910f90\") " Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.641689 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "3b02c876-c14b-4422-beaf-9ea3ab910f90" (UID: "3b02c876-c14b-4422-beaf-9ea3ab910f90"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.643434 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.648705 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31145cc8-41cc-4870-83f8-bca24095bb0c" path="/var/lib/kubelet/pods/31145cc8-41cc-4870-83f8-bca24095bb0c/volumes" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.649166 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "3b02c876-c14b-4422-beaf-9ea3ab910f90" (UID: "3b02c876-c14b-4422-beaf-9ea3ab910f90"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.649502 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" path="/var/lib/kubelet/pods/3476fe21-9207-4c63-80e9-0c3e1bd7e19c/volumes" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.650303 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="787fe453-1786-449c-bf60-87153ce058f9" path="/var/lib/kubelet/pods/787fe453-1786-449c-bf60-87153ce058f9/volumes" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.653442 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b02c876-c14b-4422-beaf-9ea3ab910f90-kube-api-access-xzrrh" (OuterVolumeSpecName: "kube-api-access-xzrrh") pod "3b02c876-c14b-4422-beaf-9ea3ab910f90" (UID: "3b02c876-c14b-4422-beaf-9ea3ab910f90"). InnerVolumeSpecName "kube-api-access-xzrrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.746594 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzrrh\" (UniqueName: \"kubernetes.io/projected/3b02c876-c14b-4422-beaf-9ea3ab910f90-kube-api-access-xzrrh\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:49 crc kubenswrapper[4799]: I1124 06:51:49.746636 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3b02c876-c14b-4422-beaf-9ea3ab910f90-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.193603 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" event={"ID":"3c598ccb-346a-45eb-ad78-ad538667a705","Type":"ContainerStarted","Data":"69d981d085cffece3a3525554059d4f0ab1af5f222b1e2850ecee474b5d2febd"} Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.193658 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" event={"ID":"3c598ccb-346a-45eb-ad78-ad538667a705","Type":"ContainerStarted","Data":"1e6f63c2c0e9d2903fdcb97a8001e8522446d1e05420d9c36892acb6be5a7a67"} Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.194179 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.198503 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzb6m" event={"ID":"7dab8ddb-4cb5-4349-a4e5-22b50fbf0740","Type":"ContainerDied","Data":"2f4b1ecb25ed5a425c4f859749db0c70af0d41e666003b74fafd0a73d84bce2d"} Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.198640 4799 scope.go:117] "RemoveContainer" containerID="ecf3144ce85fdf6e45be5912761a06823a3a942e244d8b1dd07d2dff1b728aff" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.198648 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzb6m" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.200562 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" event={"ID":"3b02c876-c14b-4422-beaf-9ea3ab910f90","Type":"ContainerDied","Data":"0455c604bc8a8efd5be8f2379d90380f01a095a7fef4d48bfc97e5e12f004f72"} Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.200604 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pcvtw" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.202359 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.215292 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2hdxh" podStartSLOduration=2.2152652059999998 podStartE2EDuration="2.215265206s" podCreationTimestamp="2025-11-24 06:51:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:51:50.212961682 +0000 UTC m=+255.868944156" watchObservedRunningTime="2025-11-24 06:51:50.215265206 +0000 UTC m=+255.871247680" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.220152 4799 scope.go:117] "RemoveContainer" containerID="d641cf589fd5a7e578892fa32d1277d6791180735fc1c1b55f40eefe15ca2f03" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.239525 4799 scope.go:117] "RemoveContainer" containerID="ed401757ec819c88507bc7c4f79a0d833d180236d8e239ca36b691815549f8ae" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.278796 4799 scope.go:117] "RemoveContainer" containerID="b4c7b3100a3fc2990475a72d5d9965c9288ff6e1f0c3c62513f32bcf494ab724" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.286998 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vzb6m"] Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.295624 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vzb6m"] Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.300357 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcvtw"] Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.304062 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pcvtw"] Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773174 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r5bgj"] Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773516 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerName="extract-utilities" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773535 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerName="extract-utilities" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773557 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" containerName="extract-content" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773568 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" containerName="extract-content" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773581 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773592 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773605 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerName="extract-content" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773615 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerName="extract-content" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773628 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="787fe453-1786-449c-bf60-87153ce058f9" containerName="extract-content" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773638 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="787fe453-1786-449c-bf60-87153ce058f9" containerName="extract-content" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773650 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31145cc8-41cc-4870-83f8-bca24095bb0c" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773661 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="31145cc8-41cc-4870-83f8-bca24095bb0c" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773678 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="787fe453-1786-449c-bf60-87153ce058f9" containerName="extract-utilities" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773687 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="787fe453-1786-449c-bf60-87153ce058f9" containerName="extract-utilities" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773697 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31145cc8-41cc-4870-83f8-bca24095bb0c" containerName="extract-content" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773706 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="31145cc8-41cc-4870-83f8-bca24095bb0c" containerName="extract-content" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773723 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="787fe453-1786-449c-bf60-87153ce058f9" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773732 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="787fe453-1786-449c-bf60-87153ce058f9" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773752 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773764 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773777 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b02c876-c14b-4422-beaf-9ea3ab910f90" containerName="marketplace-operator" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773787 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b02c876-c14b-4422-beaf-9ea3ab910f90" containerName="marketplace-operator" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773800 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" containerName="extract-utilities" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773809 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" containerName="extract-utilities" Nov 24 06:51:50 crc kubenswrapper[4799]: E1124 06:51:50.773821 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31145cc8-41cc-4870-83f8-bca24095bb0c" containerName="extract-utilities" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773832 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="31145cc8-41cc-4870-83f8-bca24095bb0c" containerName="extract-utilities" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.773987 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b02c876-c14b-4422-beaf-9ea3ab910f90" containerName="marketplace-operator" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.774009 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.774024 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="31145cc8-41cc-4870-83f8-bca24095bb0c" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.774046 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3476fe21-9207-4c63-80e9-0c3e1bd7e19c" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.774059 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="787fe453-1786-449c-bf60-87153ce058f9" containerName="registry-server" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.777761 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.781510 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.787877 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r5bgj"] Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.965328 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4wd9p"] Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.966047 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f64dad-142a-42ca-b7c0-c4adeaec75ee-utilities\") pod \"redhat-marketplace-r5bgj\" (UID: \"15f64dad-142a-42ca-b7c0-c4adeaec75ee\") " pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.966121 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgqx5\" (UniqueName: \"kubernetes.io/projected/15f64dad-142a-42ca-b7c0-c4adeaec75ee-kube-api-access-lgqx5\") pod \"redhat-marketplace-r5bgj\" (UID: \"15f64dad-142a-42ca-b7c0-c4adeaec75ee\") " pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.966182 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f64dad-142a-42ca-b7c0-c4adeaec75ee-catalog-content\") pod \"redhat-marketplace-r5bgj\" (UID: \"15f64dad-142a-42ca-b7c0-c4adeaec75ee\") " pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.968414 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.973965 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 06:51:50 crc kubenswrapper[4799]: I1124 06:51:50.977788 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4wd9p"] Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.067765 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-catalog-content\") pod \"redhat-operators-4wd9p\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.068242 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-utilities\") pod \"redhat-operators-4wd9p\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.068381 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f64dad-142a-42ca-b7c0-c4adeaec75ee-utilities\") pod \"redhat-marketplace-r5bgj\" (UID: \"15f64dad-142a-42ca-b7c0-c4adeaec75ee\") " pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.068514 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfjm4\" (UniqueName: \"kubernetes.io/projected/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-kube-api-access-lfjm4\") pod \"redhat-operators-4wd9p\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.068695 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgqx5\" (UniqueName: \"kubernetes.io/projected/15f64dad-142a-42ca-b7c0-c4adeaec75ee-kube-api-access-lgqx5\") pod \"redhat-marketplace-r5bgj\" (UID: \"15f64dad-142a-42ca-b7c0-c4adeaec75ee\") " pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.068893 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f64dad-142a-42ca-b7c0-c4adeaec75ee-catalog-content\") pod \"redhat-marketplace-r5bgj\" (UID: \"15f64dad-142a-42ca-b7c0-c4adeaec75ee\") " pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.068935 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f64dad-142a-42ca-b7c0-c4adeaec75ee-utilities\") pod \"redhat-marketplace-r5bgj\" (UID: \"15f64dad-142a-42ca-b7c0-c4adeaec75ee\") " pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.069287 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f64dad-142a-42ca-b7c0-c4adeaec75ee-catalog-content\") pod \"redhat-marketplace-r5bgj\" (UID: \"15f64dad-142a-42ca-b7c0-c4adeaec75ee\") " pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.095869 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgqx5\" (UniqueName: \"kubernetes.io/projected/15f64dad-142a-42ca-b7c0-c4adeaec75ee-kube-api-access-lgqx5\") pod \"redhat-marketplace-r5bgj\" (UID: \"15f64dad-142a-42ca-b7c0-c4adeaec75ee\") " pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.097480 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.171080 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfjm4\" (UniqueName: \"kubernetes.io/projected/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-kube-api-access-lfjm4\") pod \"redhat-operators-4wd9p\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.171198 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-catalog-content\") pod \"redhat-operators-4wd9p\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.171245 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-utilities\") pod \"redhat-operators-4wd9p\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.171904 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-catalog-content\") pod \"redhat-operators-4wd9p\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.171966 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-utilities\") pod \"redhat-operators-4wd9p\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.198994 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfjm4\" (UniqueName: \"kubernetes.io/projected/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-kube-api-access-lfjm4\") pod \"redhat-operators-4wd9p\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.288202 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.512010 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r5bgj"] Nov 24 06:51:51 crc kubenswrapper[4799]: W1124 06:51:51.522293 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15f64dad_142a_42ca_b7c0_c4adeaec75ee.slice/crio-f1b7735d4493656ad681b8d7170ae5b3f7b6ecced5942c22a97a3ffa8143a9fb WatchSource:0}: Error finding container f1b7735d4493656ad681b8d7170ae5b3f7b6ecced5942c22a97a3ffa8143a9fb: Status 404 returned error can't find the container with id f1b7735d4493656ad681b8d7170ae5b3f7b6ecced5942c22a97a3ffa8143a9fb Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.636546 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b02c876-c14b-4422-beaf-9ea3ab910f90" path="/var/lib/kubelet/pods/3b02c876-c14b-4422-beaf-9ea3ab910f90/volumes" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.637120 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dab8ddb-4cb5-4349-a4e5-22b50fbf0740" path="/var/lib/kubelet/pods/7dab8ddb-4cb5-4349-a4e5-22b50fbf0740/volumes" Nov 24 06:51:51 crc kubenswrapper[4799]: I1124 06:51:51.686385 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4wd9p"] Nov 24 06:51:51 crc kubenswrapper[4799]: W1124 06:51:51.752149 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode48bbcd3_6bb0_4f7d_9bac_46dd1a70e624.slice/crio-d9c21babf0e7af128a8de425d8d68dd526f65d3c5b32e60bb184e4186ddf6b31 WatchSource:0}: Error finding container d9c21babf0e7af128a8de425d8d68dd526f65d3c5b32e60bb184e4186ddf6b31: Status 404 returned error can't find the container with id d9c21babf0e7af128a8de425d8d68dd526f65d3c5b32e60bb184e4186ddf6b31 Nov 24 06:51:52 crc kubenswrapper[4799]: I1124 06:51:52.226126 4799 generic.go:334] "Generic (PLEG): container finished" podID="15f64dad-142a-42ca-b7c0-c4adeaec75ee" containerID="39affea6803a77c11eea05dd487d4c4493a4e6aec9de6b7b77b00bbd78bc7068" exitCode=0 Nov 24 06:51:52 crc kubenswrapper[4799]: I1124 06:51:52.226336 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r5bgj" event={"ID":"15f64dad-142a-42ca-b7c0-c4adeaec75ee","Type":"ContainerDied","Data":"39affea6803a77c11eea05dd487d4c4493a4e6aec9de6b7b77b00bbd78bc7068"} Nov 24 06:51:52 crc kubenswrapper[4799]: I1124 06:51:52.227024 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r5bgj" event={"ID":"15f64dad-142a-42ca-b7c0-c4adeaec75ee","Type":"ContainerStarted","Data":"f1b7735d4493656ad681b8d7170ae5b3f7b6ecced5942c22a97a3ffa8143a9fb"} Nov 24 06:51:52 crc kubenswrapper[4799]: I1124 06:51:52.231418 4799 generic.go:334] "Generic (PLEG): container finished" podID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" containerID="3d4549c36635a0a490a1efb8b0bea23c5a2b35e2f2b47bb44afc033260334f85" exitCode=0 Nov 24 06:51:52 crc kubenswrapper[4799]: I1124 06:51:52.231535 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wd9p" event={"ID":"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624","Type":"ContainerDied","Data":"3d4549c36635a0a490a1efb8b0bea23c5a2b35e2f2b47bb44afc033260334f85"} Nov 24 06:51:52 crc kubenswrapper[4799]: I1124 06:51:52.231590 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wd9p" event={"ID":"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624","Type":"ContainerStarted","Data":"d9c21babf0e7af128a8de425d8d68dd526f65d3c5b32e60bb184e4186ddf6b31"} Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.158622 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s2vvw"] Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.160331 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.164348 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.176834 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s2vvw"] Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.239304 4799 generic.go:334] "Generic (PLEG): container finished" podID="15f64dad-142a-42ca-b7c0-c4adeaec75ee" containerID="60450551c57a18dba0413632892853926b0f2e1ec00cc3f5faf42afd90a71045" exitCode=0 Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.239396 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r5bgj" event={"ID":"15f64dad-142a-42ca-b7c0-c4adeaec75ee","Type":"ContainerDied","Data":"60450551c57a18dba0413632892853926b0f2e1ec00cc3f5faf42afd90a71045"} Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.242456 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wd9p" event={"ID":"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624","Type":"ContainerStarted","Data":"900bb6ba8a0319762fe7b1b5c9b3f72e1bc6a74964d4915427b7ec55ba29ed95"} Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.304635 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w86t7\" (UniqueName: \"kubernetes.io/projected/b7277667-0bca-4cb2-9e44-304597d405c8-kube-api-access-w86t7\") pod \"certified-operators-s2vvw\" (UID: \"b7277667-0bca-4cb2-9e44-304597d405c8\") " pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.304729 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7277667-0bca-4cb2-9e44-304597d405c8-catalog-content\") pod \"certified-operators-s2vvw\" (UID: \"b7277667-0bca-4cb2-9e44-304597d405c8\") " pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.304752 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7277667-0bca-4cb2-9e44-304597d405c8-utilities\") pod \"certified-operators-s2vvw\" (UID: \"b7277667-0bca-4cb2-9e44-304597d405c8\") " pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.360940 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9spwl"] Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.362305 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.365405 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.375077 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9spwl"] Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.406017 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7277667-0bca-4cb2-9e44-304597d405c8-catalog-content\") pod \"certified-operators-s2vvw\" (UID: \"b7277667-0bca-4cb2-9e44-304597d405c8\") " pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.406079 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7277667-0bca-4cb2-9e44-304597d405c8-utilities\") pod \"certified-operators-s2vvw\" (UID: \"b7277667-0bca-4cb2-9e44-304597d405c8\") " pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.406133 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w86t7\" (UniqueName: \"kubernetes.io/projected/b7277667-0bca-4cb2-9e44-304597d405c8-kube-api-access-w86t7\") pod \"certified-operators-s2vvw\" (UID: \"b7277667-0bca-4cb2-9e44-304597d405c8\") " pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.406998 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7277667-0bca-4cb2-9e44-304597d405c8-catalog-content\") pod \"certified-operators-s2vvw\" (UID: \"b7277667-0bca-4cb2-9e44-304597d405c8\") " pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.407718 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7277667-0bca-4cb2-9e44-304597d405c8-utilities\") pod \"certified-operators-s2vvw\" (UID: \"b7277667-0bca-4cb2-9e44-304597d405c8\") " pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.437164 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w86t7\" (UniqueName: \"kubernetes.io/projected/b7277667-0bca-4cb2-9e44-304597d405c8-kube-api-access-w86t7\") pod \"certified-operators-s2vvw\" (UID: \"b7277667-0bca-4cb2-9e44-304597d405c8\") " pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.481754 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.507786 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skgjt\" (UniqueName: \"kubernetes.io/projected/1282ec5d-e64d-463c-b5b3-15f41587f498-kube-api-access-skgjt\") pod \"community-operators-9spwl\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.507872 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-catalog-content\") pod \"community-operators-9spwl\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.507950 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-utilities\") pod \"community-operators-9spwl\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.609730 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-utilities\") pod \"community-operators-9spwl\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.611228 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-utilities\") pod \"community-operators-9spwl\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.611468 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skgjt\" (UniqueName: \"kubernetes.io/projected/1282ec5d-e64d-463c-b5b3-15f41587f498-kube-api-access-skgjt\") pod \"community-operators-9spwl\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.611673 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-catalog-content\") pod \"community-operators-9spwl\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.612059 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-catalog-content\") pod \"community-operators-9spwl\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.636646 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skgjt\" (UniqueName: \"kubernetes.io/projected/1282ec5d-e64d-463c-b5b3-15f41587f498-kube-api-access-skgjt\") pod \"community-operators-9spwl\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.679513 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.899106 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9spwl"] Nov 24 06:51:53 crc kubenswrapper[4799]: W1124 06:51:53.909057 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1282ec5d_e64d_463c_b5b3_15f41587f498.slice/crio-5cebebcef9944ef817943117d07c981215a71c1a8c4bf6f78ca08d4dcd809a5e WatchSource:0}: Error finding container 5cebebcef9944ef817943117d07c981215a71c1a8c4bf6f78ca08d4dcd809a5e: Status 404 returned error can't find the container with id 5cebebcef9944ef817943117d07c981215a71c1a8c4bf6f78ca08d4dcd809a5e Nov 24 06:51:53 crc kubenswrapper[4799]: I1124 06:51:53.938420 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s2vvw"] Nov 24 06:51:53 crc kubenswrapper[4799]: W1124 06:51:53.951472 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7277667_0bca_4cb2_9e44_304597d405c8.slice/crio-1b0d338bf531ac0b230d1a81c555e63e71090f7bd870ffbe9ae408ce536bdae1 WatchSource:0}: Error finding container 1b0d338bf531ac0b230d1a81c555e63e71090f7bd870ffbe9ae408ce536bdae1: Status 404 returned error can't find the container with id 1b0d338bf531ac0b230d1a81c555e63e71090f7bd870ffbe9ae408ce536bdae1 Nov 24 06:51:54 crc kubenswrapper[4799]: I1124 06:51:54.250645 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r5bgj" event={"ID":"15f64dad-142a-42ca-b7c0-c4adeaec75ee","Type":"ContainerStarted","Data":"cf1d31c5a74c9b1d330602d7f9f496ab32a9ecc54e36cc4c561d39eaee937250"} Nov 24 06:51:54 crc kubenswrapper[4799]: I1124 06:51:54.253525 4799 generic.go:334] "Generic (PLEG): container finished" podID="1282ec5d-e64d-463c-b5b3-15f41587f498" containerID="d142866ab8960171954d38ac87bbcf6697a0fe5151d50fe1a0460987cf823424" exitCode=0 Nov 24 06:51:54 crc kubenswrapper[4799]: I1124 06:51:54.253604 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9spwl" event={"ID":"1282ec5d-e64d-463c-b5b3-15f41587f498","Type":"ContainerDied","Data":"d142866ab8960171954d38ac87bbcf6697a0fe5151d50fe1a0460987cf823424"} Nov 24 06:51:54 crc kubenswrapper[4799]: I1124 06:51:54.253640 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9spwl" event={"ID":"1282ec5d-e64d-463c-b5b3-15f41587f498","Type":"ContainerStarted","Data":"5cebebcef9944ef817943117d07c981215a71c1a8c4bf6f78ca08d4dcd809a5e"} Nov 24 06:51:54 crc kubenswrapper[4799]: I1124 06:51:54.256834 4799 generic.go:334] "Generic (PLEG): container finished" podID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" containerID="900bb6ba8a0319762fe7b1b5c9b3f72e1bc6a74964d4915427b7ec55ba29ed95" exitCode=0 Nov 24 06:51:54 crc kubenswrapper[4799]: I1124 06:51:54.256895 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wd9p" event={"ID":"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624","Type":"ContainerDied","Data":"900bb6ba8a0319762fe7b1b5c9b3f72e1bc6a74964d4915427b7ec55ba29ed95"} Nov 24 06:51:54 crc kubenswrapper[4799]: I1124 06:51:54.258568 4799 generic.go:334] "Generic (PLEG): container finished" podID="b7277667-0bca-4cb2-9e44-304597d405c8" containerID="002b731417ae8e1c810572572a6dfaf706c5ffc6bf7d086b06e69b5fff331533" exitCode=0 Nov 24 06:51:54 crc kubenswrapper[4799]: I1124 06:51:54.258595 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2vvw" event={"ID":"b7277667-0bca-4cb2-9e44-304597d405c8","Type":"ContainerDied","Data":"002b731417ae8e1c810572572a6dfaf706c5ffc6bf7d086b06e69b5fff331533"} Nov 24 06:51:54 crc kubenswrapper[4799]: I1124 06:51:54.258610 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2vvw" event={"ID":"b7277667-0bca-4cb2-9e44-304597d405c8","Type":"ContainerStarted","Data":"1b0d338bf531ac0b230d1a81c555e63e71090f7bd870ffbe9ae408ce536bdae1"} Nov 24 06:51:54 crc kubenswrapper[4799]: I1124 06:51:54.300650 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r5bgj" podStartSLOduration=2.8173174039999997 podStartE2EDuration="4.300620717s" podCreationTimestamp="2025-11-24 06:51:50 +0000 UTC" firstStartedPulling="2025-11-24 06:51:52.229011726 +0000 UTC m=+257.884994210" lastFinishedPulling="2025-11-24 06:51:53.712315049 +0000 UTC m=+259.368297523" observedRunningTime="2025-11-24 06:51:54.274458526 +0000 UTC m=+259.930441000" watchObservedRunningTime="2025-11-24 06:51:54.300620717 +0000 UTC m=+259.956603191" Nov 24 06:51:55 crc kubenswrapper[4799]: I1124 06:51:55.285728 4799 generic.go:334] "Generic (PLEG): container finished" podID="1282ec5d-e64d-463c-b5b3-15f41587f498" containerID="a66a7fda1ec8146d70589bd8eb6e793b7edeed60e7b21ba935ad2d36a29e4abb" exitCode=0 Nov 24 06:51:55 crc kubenswrapper[4799]: I1124 06:51:55.286214 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9spwl" event={"ID":"1282ec5d-e64d-463c-b5b3-15f41587f498","Type":"ContainerDied","Data":"a66a7fda1ec8146d70589bd8eb6e793b7edeed60e7b21ba935ad2d36a29e4abb"} Nov 24 06:51:55 crc kubenswrapper[4799]: I1124 06:51:55.292769 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wd9p" event={"ID":"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624","Type":"ContainerStarted","Data":"ec4e4bc1527a4dbc22f991c1ed8d40c3f2b359f4bfbb1fe649149bf37ab7f83b"} Nov 24 06:51:55 crc kubenswrapper[4799]: I1124 06:51:55.296115 4799 generic.go:334] "Generic (PLEG): container finished" podID="b7277667-0bca-4cb2-9e44-304597d405c8" containerID="507e37465aa65c6d9dbf17d00abeda8bbbf793e88467ec51b35fd8ad4256abb6" exitCode=0 Nov 24 06:51:55 crc kubenswrapper[4799]: I1124 06:51:55.296286 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2vvw" event={"ID":"b7277667-0bca-4cb2-9e44-304597d405c8","Type":"ContainerDied","Data":"507e37465aa65c6d9dbf17d00abeda8bbbf793e88467ec51b35fd8ad4256abb6"} Nov 24 06:51:55 crc kubenswrapper[4799]: I1124 06:51:55.344546 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4wd9p" podStartSLOduration=2.935577455 podStartE2EDuration="5.344513426s" podCreationTimestamp="2025-11-24 06:51:50 +0000 UTC" firstStartedPulling="2025-11-24 06:51:52.237446937 +0000 UTC m=+257.893429411" lastFinishedPulling="2025-11-24 06:51:54.646382908 +0000 UTC m=+260.302365382" observedRunningTime="2025-11-24 06:51:55.328549663 +0000 UTC m=+260.984532147" watchObservedRunningTime="2025-11-24 06:51:55.344513426 +0000 UTC m=+261.000495900" Nov 24 06:51:56 crc kubenswrapper[4799]: I1124 06:51:56.313259 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9spwl" event={"ID":"1282ec5d-e64d-463c-b5b3-15f41587f498","Type":"ContainerStarted","Data":"94126a15a7e912bac3e6b31ea4c2ced1f6c1dc91b7493ef32a31b262c15a8f90"} Nov 24 06:51:56 crc kubenswrapper[4799]: I1124 06:51:56.316202 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2vvw" event={"ID":"b7277667-0bca-4cb2-9e44-304597d405c8","Type":"ContainerStarted","Data":"906c0d7014f51045028a5da4fe853d56994021ec5712e48b5fabaaee9eeca1bf"} Nov 24 06:51:56 crc kubenswrapper[4799]: I1124 06:51:56.333507 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9spwl" podStartSLOduration=1.913667846 podStartE2EDuration="3.333490007s" podCreationTimestamp="2025-11-24 06:51:53 +0000 UTC" firstStartedPulling="2025-11-24 06:51:54.255386033 +0000 UTC m=+259.911368507" lastFinishedPulling="2025-11-24 06:51:55.675208194 +0000 UTC m=+261.331190668" observedRunningTime="2025-11-24 06:51:56.329432515 +0000 UTC m=+261.985414989" watchObservedRunningTime="2025-11-24 06:51:56.333490007 +0000 UTC m=+261.989472471" Nov 24 06:51:56 crc kubenswrapper[4799]: I1124 06:51:56.353941 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s2vvw" podStartSLOduration=1.926702201 podStartE2EDuration="3.35391373s" podCreationTimestamp="2025-11-24 06:51:53 +0000 UTC" firstStartedPulling="2025-11-24 06:51:54.259734363 +0000 UTC m=+259.915716837" lastFinishedPulling="2025-11-24 06:51:55.686945892 +0000 UTC m=+261.342928366" observedRunningTime="2025-11-24 06:51:56.351518468 +0000 UTC m=+262.007500952" watchObservedRunningTime="2025-11-24 06:51:56.35391373 +0000 UTC m=+262.009896204" Nov 24 06:52:01 crc kubenswrapper[4799]: I1124 06:52:01.098356 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:52:01 crc kubenswrapper[4799]: I1124 06:52:01.099372 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:52:01 crc kubenswrapper[4799]: I1124 06:52:01.151516 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:52:01 crc kubenswrapper[4799]: I1124 06:52:01.294037 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:52:01 crc kubenswrapper[4799]: I1124 06:52:01.294989 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:52:01 crc kubenswrapper[4799]: I1124 06:52:01.359506 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:52:01 crc kubenswrapper[4799]: I1124 06:52:01.396144 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r5bgj" Nov 24 06:52:02 crc kubenswrapper[4799]: I1124 06:52:02.395450 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 06:52:03 crc kubenswrapper[4799]: I1124 06:52:03.482112 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:52:03 crc kubenswrapper[4799]: I1124 06:52:03.483119 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:52:03 crc kubenswrapper[4799]: I1124 06:52:03.526730 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:52:03 crc kubenswrapper[4799]: I1124 06:52:03.680490 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:52:03 crc kubenswrapper[4799]: I1124 06:52:03.680841 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:52:03 crc kubenswrapper[4799]: I1124 06:52:03.726726 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:52:04 crc kubenswrapper[4799]: I1124 06:52:04.402916 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s2vvw" Nov 24 06:52:04 crc kubenswrapper[4799]: I1124 06:52:04.428273 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9spwl" Nov 24 06:53:20 crc kubenswrapper[4799]: I1124 06:53:20.401346 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:53:20 crc kubenswrapper[4799]: I1124 06:53:20.402506 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:53:50 crc kubenswrapper[4799]: I1124 06:53:50.400778 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:53:50 crc kubenswrapper[4799]: I1124 06:53:50.403699 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:54:20 crc kubenswrapper[4799]: I1124 06:54:20.400298 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:54:20 crc kubenswrapper[4799]: I1124 06:54:20.401236 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:54:20 crc kubenswrapper[4799]: I1124 06:54:20.401385 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:54:20 crc kubenswrapper[4799]: I1124 06:54:20.402323 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"81921defcb796b9a98513b3ce531e00df43592241cc2ef5d7a1b9452de61e2c8"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 06:54:20 crc kubenswrapper[4799]: I1124 06:54:20.402410 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://81921defcb796b9a98513b3ce531e00df43592241cc2ef5d7a1b9452de61e2c8" gracePeriod=600 Nov 24 06:54:20 crc kubenswrapper[4799]: E1124 06:54:20.479678 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod354ebd50_7a67_481d_86db_4b3bf4753161.slice/crio-conmon-81921defcb796b9a98513b3ce531e00df43592241cc2ef5d7a1b9452de61e2c8.scope\": RecentStats: unable to find data in memory cache]" Nov 24 06:54:21 crc kubenswrapper[4799]: I1124 06:54:21.371154 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="81921defcb796b9a98513b3ce531e00df43592241cc2ef5d7a1b9452de61e2c8" exitCode=0 Nov 24 06:54:21 crc kubenswrapper[4799]: I1124 06:54:21.372216 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"81921defcb796b9a98513b3ce531e00df43592241cc2ef5d7a1b9452de61e2c8"} Nov 24 06:54:21 crc kubenswrapper[4799]: I1124 06:54:21.372270 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"1155080c324c74d531908515f79c186dcf0bbbcf9d2021dbf86a19ed644ed48b"} Nov 24 06:54:21 crc kubenswrapper[4799]: I1124 06:54:21.372302 4799 scope.go:117] "RemoveContainer" containerID="271d61702e4849eb10b435fcca30f224a1dd78390f312b21cc25ca74601c7df9" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.600936 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wxfqr"] Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.603977 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.618210 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wxfqr"] Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.793152 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-registry-tls\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.793214 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nbxs\" (UniqueName: \"kubernetes.io/projected/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-kube-api-access-2nbxs\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.793281 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-registry-certificates\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.793488 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.793557 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-trusted-ca\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.793616 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.793697 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.793738 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-bound-sa-token\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.815440 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.908444 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-bound-sa-token\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.908542 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-registry-tls\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.908627 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nbxs\" (UniqueName: \"kubernetes.io/projected/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-kube-api-access-2nbxs\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.908719 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-registry-certificates\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.908768 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.908797 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-trusted-ca\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.908860 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.910950 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.911558 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-registry-certificates\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.913387 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-trusted-ca\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.920889 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.924931 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-registry-tls\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.927828 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-bound-sa-token\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:09 crc kubenswrapper[4799]: I1124 06:55:09.928315 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nbxs\" (UniqueName: \"kubernetes.io/projected/67a2c10b-9751-4dde-aa6b-9b4469b4f28e-kube-api-access-2nbxs\") pod \"image-registry-66df7c8f76-wxfqr\" (UID: \"67a2c10b-9751-4dde-aa6b-9b4469b4f28e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:10 crc kubenswrapper[4799]: I1124 06:55:10.222435 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:10 crc kubenswrapper[4799]: I1124 06:55:10.432079 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wxfqr"] Nov 24 06:55:10 crc kubenswrapper[4799]: I1124 06:55:10.733695 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" event={"ID":"67a2c10b-9751-4dde-aa6b-9b4469b4f28e","Type":"ContainerStarted","Data":"71ccbc1cf1a80f8801154a3677d16e742902398adbbb69b002626f3f1df77219"} Nov 24 06:55:10 crc kubenswrapper[4799]: I1124 06:55:10.734304 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:10 crc kubenswrapper[4799]: I1124 06:55:10.734326 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" event={"ID":"67a2c10b-9751-4dde-aa6b-9b4469b4f28e","Type":"ContainerStarted","Data":"58bc1e6a1f40db2b249baf783bac8fdc0df3fab99c2588cb36004267026d0c2d"} Nov 24 06:55:30 crc kubenswrapper[4799]: I1124 06:55:30.229240 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" Nov 24 06:55:30 crc kubenswrapper[4799]: I1124 06:55:30.261136 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-wxfqr" podStartSLOduration=21.26110652 podStartE2EDuration="21.26110652s" podCreationTimestamp="2025-11-24 06:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:55:10.763696873 +0000 UTC m=+456.419679367" watchObservedRunningTime="2025-11-24 06:55:30.26110652 +0000 UTC m=+475.917088994" Nov 24 06:55:30 crc kubenswrapper[4799]: I1124 06:55:30.299269 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5bjhn"] Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.357645 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" podUID="b7761980-8663-4dda-924b-2fb787fcdac8" containerName="registry" containerID="cri-o://c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1" gracePeriod=30 Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.785012 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.845000 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-bound-sa-token\") pod \"b7761980-8663-4dda-924b-2fb787fcdac8\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.845068 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b7761980-8663-4dda-924b-2fb787fcdac8-installation-pull-secrets\") pod \"b7761980-8663-4dda-924b-2fb787fcdac8\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.845103 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-trusted-ca\") pod \"b7761980-8663-4dda-924b-2fb787fcdac8\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.845139 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b7761980-8663-4dda-924b-2fb787fcdac8-ca-trust-extracted\") pod \"b7761980-8663-4dda-924b-2fb787fcdac8\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.845161 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-registry-tls\") pod \"b7761980-8663-4dda-924b-2fb787fcdac8\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.845308 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"b7761980-8663-4dda-924b-2fb787fcdac8\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.845359 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtzdz\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-kube-api-access-gtzdz\") pod \"b7761980-8663-4dda-924b-2fb787fcdac8\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.845392 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-registry-certificates\") pod \"b7761980-8663-4dda-924b-2fb787fcdac8\" (UID: \"b7761980-8663-4dda-924b-2fb787fcdac8\") " Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.846528 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "b7761980-8663-4dda-924b-2fb787fcdac8" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.847307 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "b7761980-8663-4dda-924b-2fb787fcdac8" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.854717 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-kube-api-access-gtzdz" (OuterVolumeSpecName: "kube-api-access-gtzdz") pod "b7761980-8663-4dda-924b-2fb787fcdac8" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8"). InnerVolumeSpecName "kube-api-access-gtzdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.855362 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "b7761980-8663-4dda-924b-2fb787fcdac8" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.856323 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7761980-8663-4dda-924b-2fb787fcdac8-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "b7761980-8663-4dda-924b-2fb787fcdac8" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.856537 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "b7761980-8663-4dda-924b-2fb787fcdac8" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.863633 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "b7761980-8663-4dda-924b-2fb787fcdac8" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.869907 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7761980-8663-4dda-924b-2fb787fcdac8-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "b7761980-8663-4dda-924b-2fb787fcdac8" (UID: "b7761980-8663-4dda-924b-2fb787fcdac8"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.947043 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtzdz\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-kube-api-access-gtzdz\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.947087 4799 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.947098 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.947107 4799 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b7761980-8663-4dda-924b-2fb787fcdac8-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.947121 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7761980-8663-4dda-924b-2fb787fcdac8-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.947131 4799 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b7761980-8663-4dda-924b-2fb787fcdac8-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:55 crc kubenswrapper[4799]: I1124 06:55:55.947139 4799 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b7761980-8663-4dda-924b-2fb787fcdac8-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 06:55:56 crc kubenswrapper[4799]: I1124 06:55:56.065657 4799 generic.go:334] "Generic (PLEG): container finished" podID="b7761980-8663-4dda-924b-2fb787fcdac8" containerID="c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1" exitCode=0 Nov 24 06:55:56 crc kubenswrapper[4799]: I1124 06:55:56.065719 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" event={"ID":"b7761980-8663-4dda-924b-2fb787fcdac8","Type":"ContainerDied","Data":"c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1"} Nov 24 06:55:56 crc kubenswrapper[4799]: I1124 06:55:56.065753 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" event={"ID":"b7761980-8663-4dda-924b-2fb787fcdac8","Type":"ContainerDied","Data":"78e78ea44ec9153887432e0a140338778ae9d10337a0c152ec7319ba64863063"} Nov 24 06:55:56 crc kubenswrapper[4799]: I1124 06:55:56.065756 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5bjhn" Nov 24 06:55:56 crc kubenswrapper[4799]: I1124 06:55:56.065793 4799 scope.go:117] "RemoveContainer" containerID="c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1" Nov 24 06:55:56 crc kubenswrapper[4799]: I1124 06:55:56.088952 4799 scope.go:117] "RemoveContainer" containerID="c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1" Nov 24 06:55:56 crc kubenswrapper[4799]: E1124 06:55:56.089703 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1\": container with ID starting with c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1 not found: ID does not exist" containerID="c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1" Nov 24 06:55:56 crc kubenswrapper[4799]: I1124 06:55:56.089762 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1"} err="failed to get container status \"c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1\": rpc error: code = NotFound desc = could not find container \"c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1\": container with ID starting with c403d32c8f6572a679f26b6528069736b92054cc57aad06fbc4233ce31723bd1 not found: ID does not exist" Nov 24 06:55:56 crc kubenswrapper[4799]: I1124 06:55:56.099029 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5bjhn"] Nov 24 06:55:56 crc kubenswrapper[4799]: I1124 06:55:56.101893 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5bjhn"] Nov 24 06:55:57 crc kubenswrapper[4799]: I1124 06:55:57.640749 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7761980-8663-4dda-924b-2fb787fcdac8" path="/var/lib/kubelet/pods/b7761980-8663-4dda-924b-2fb787fcdac8/volumes" Nov 24 06:56:20 crc kubenswrapper[4799]: I1124 06:56:20.400914 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:56:20 crc kubenswrapper[4799]: I1124 06:56:20.402015 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:56:50 crc kubenswrapper[4799]: I1124 06:56:50.400826 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:56:50 crc kubenswrapper[4799]: I1124 06:56:50.401998 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:57:20 crc kubenswrapper[4799]: I1124 06:57:20.401367 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:57:20 crc kubenswrapper[4799]: I1124 06:57:20.402271 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:57:20 crc kubenswrapper[4799]: I1124 06:57:20.402363 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 06:57:20 crc kubenswrapper[4799]: I1124 06:57:20.403349 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1155080c324c74d531908515f79c186dcf0bbbcf9d2021dbf86a19ed644ed48b"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 06:57:20 crc kubenswrapper[4799]: I1124 06:57:20.403490 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://1155080c324c74d531908515f79c186dcf0bbbcf9d2021dbf86a19ed644ed48b" gracePeriod=600 Nov 24 06:57:20 crc kubenswrapper[4799]: I1124 06:57:20.663351 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="1155080c324c74d531908515f79c186dcf0bbbcf9d2021dbf86a19ed644ed48b" exitCode=0 Nov 24 06:57:20 crc kubenswrapper[4799]: I1124 06:57:20.663428 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"1155080c324c74d531908515f79c186dcf0bbbcf9d2021dbf86a19ed644ed48b"} Nov 24 06:57:20 crc kubenswrapper[4799]: I1124 06:57:20.664051 4799 scope.go:117] "RemoveContainer" containerID="81921defcb796b9a98513b3ce531e00df43592241cc2ef5d7a1b9452de61e2c8" Nov 24 06:57:21 crc kubenswrapper[4799]: I1124 06:57:21.675625 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"d29a0547b29f459c7d7db9037c62a8945e7f6c626531ca09a0a9fd2c39286a5a"} Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.255613 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b7nd7"] Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.257580 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovn-controller" containerID="cri-o://4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a" gracePeriod=30 Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.257802 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="nbdb" containerID="cri-o://85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95" gracePeriod=30 Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.257916 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="sbdb" containerID="cri-o://28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54" gracePeriod=30 Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.257968 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="kube-rbac-proxy-node" containerID="cri-o://ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74" gracePeriod=30 Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.258061 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="northd" containerID="cri-o://27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1" gracePeriod=30 Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.258091 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a" gracePeriod=30 Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.257992 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovn-acl-logging" containerID="cri-o://90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a" gracePeriod=30 Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.317205 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" containerID="cri-o://f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2" gracePeriod=30 Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.622706 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/3.log" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.625096 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovn-acl-logging/0.log" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.625533 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovn-controller/0.log" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.626516 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.697426 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fwwmx"] Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.698730 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.698778 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699226 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699239 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699253 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699261 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699272 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699279 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699290 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovn-acl-logging" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699299 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovn-acl-logging" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699311 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="kubecfg-setup" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699320 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="kubecfg-setup" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699333 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699341 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699351 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="northd" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699359 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="northd" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699371 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovn-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699379 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovn-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699388 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="nbdb" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699394 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="nbdb" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699405 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="sbdb" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699413 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="sbdb" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699425 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7761980-8663-4dda-924b-2fb787fcdac8" containerName="registry" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699432 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7761980-8663-4dda-924b-2fb787fcdac8" containerName="registry" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.699445 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="kube-rbac-proxy-node" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699468 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="kube-rbac-proxy-node" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699817 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699838 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699873 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="kube-rbac-proxy-node" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699884 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="sbdb" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699897 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovn-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699907 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="northd" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699915 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699922 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovn-acl-logging" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699929 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699964 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="nbdb" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.699975 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7761980-8663-4dda-924b-2fb787fcdac8" containerName="registry" Nov 24 06:58:15 crc kubenswrapper[4799]: E1124 06:58:15.700081 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.700088 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.700208 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.700217 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerName="ovnkube-controller" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.704910 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.760822 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-netd\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.760931 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-openvswitch\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.760970 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-ovn\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761011 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-var-lib-openvswitch\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761064 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-var-lib-cni-networks-ovn-kubernetes\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761095 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-node-log\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761147 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-kubelet\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761197 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-config\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761242 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovn-node-metrics-cert\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761286 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8djp\" (UniqueName: \"kubernetes.io/projected/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-kube-api-access-r8djp\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761328 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-netns\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761377 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-systemd\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761425 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-bin\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761477 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-etc-openvswitch\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761540 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-script-lib\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761617 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-env-overrides\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761648 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-ovn-kubernetes\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761679 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-systemd-units\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761716 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-log-socket\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.761757 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-slash\") pod \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\" (UID: \"5c34c957-e45b-4c65-8d6a-60e0a45b2f25\") " Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.762216 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-slash" (OuterVolumeSpecName: "host-slash") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.762290 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.762329 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.762367 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.762404 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.762443 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.762480 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-node-log" (OuterVolumeSpecName: "node-log") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.762515 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.763371 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.764917 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.764984 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.765011 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-log-socket" (OuterVolumeSpecName: "log-socket") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.764788 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.765279 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.765393 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.765504 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.765514 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.771140 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-kube-api-access-r8djp" (OuterVolumeSpecName: "kube-api-access-r8djp") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "kube-api-access-r8djp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.771598 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.781745 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "5c34c957-e45b-4c65-8d6a-60e0a45b2f25" (UID: "5c34c957-e45b-4c65-8d6a-60e0a45b2f25"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863118 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-var-lib-openvswitch\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863197 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-slash\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863231 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/64c312a9-19c5-4feb-85e5-95c80da1e647-ovn-node-metrics-cert\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863259 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/64c312a9-19c5-4feb-85e5-95c80da1e647-env-overrides\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863284 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-systemd-units\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863456 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-log-socket\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863549 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863586 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/64c312a9-19c5-4feb-85e5-95c80da1e647-ovnkube-config\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863615 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-run-netns\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863647 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-run-ovn\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863671 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-cni-netd\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863749 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-run-systemd\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863794 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-run-openvswitch\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863824 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/64c312a9-19c5-4feb-85e5-95c80da1e647-ovnkube-script-lib\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863881 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9jhx\" (UniqueName: \"kubernetes.io/projected/64c312a9-19c5-4feb-85e5-95c80da1e647-kube-api-access-t9jhx\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863918 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-etc-openvswitch\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863943 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-run-ovn-kubernetes\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.863967 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-node-log\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864040 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-kubelet\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864104 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-cni-bin\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864569 4799 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864599 4799 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864614 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864630 4799 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864646 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864659 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8djp\" (UniqueName: \"kubernetes.io/projected/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-kube-api-access-r8djp\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864672 4799 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864687 4799 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864703 4799 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864715 4799 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864728 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864739 4799 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864751 4799 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864764 4799 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864776 4799 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864788 4799 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864800 4799 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864814 4799 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864828 4799 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.864862 4799 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c34c957-e45b-4c65-8d6a-60e0a45b2f25-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966023 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-slash\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966153 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/64c312a9-19c5-4feb-85e5-95c80da1e647-ovn-node-metrics-cert\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966198 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/64c312a9-19c5-4feb-85e5-95c80da1e647-env-overrides\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966218 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-slash\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966237 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-systemd-units\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966273 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-log-socket\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966351 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966387 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/64c312a9-19c5-4feb-85e5-95c80da1e647-ovnkube-config\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966419 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-cni-netd\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966452 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-run-netns\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966481 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-run-ovn\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966518 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-run-systemd\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966552 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-run-openvswitch\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966585 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9jhx\" (UniqueName: \"kubernetes.io/projected/64c312a9-19c5-4feb-85e5-95c80da1e647-kube-api-access-t9jhx\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966615 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/64c312a9-19c5-4feb-85e5-95c80da1e647-ovnkube-script-lib\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966647 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-etc-openvswitch\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966713 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-run-ovn-kubernetes\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966754 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-node-log\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966801 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-kubelet\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966890 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-cni-bin\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.966963 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-var-lib-openvswitch\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967026 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-run-systemd\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967094 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-run-openvswitch\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967061 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-var-lib-openvswitch\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967158 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-systemd-units\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967190 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-log-socket\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967221 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967234 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/64c312a9-19c5-4feb-85e5-95c80da1e647-env-overrides\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967337 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-run-ovn-kubernetes\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967657 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-node-log\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967712 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-kubelet\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967757 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-cni-bin\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967801 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-etc-openvswitch\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967869 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-run-netns\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967916 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-host-cni-netd\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967958 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/64c312a9-19c5-4feb-85e5-95c80da1e647-run-ovn\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.967988 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/64c312a9-19c5-4feb-85e5-95c80da1e647-ovnkube-config\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.968298 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/64c312a9-19c5-4feb-85e5-95c80da1e647-ovnkube-script-lib\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.972059 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/64c312a9-19c5-4feb-85e5-95c80da1e647-ovn-node-metrics-cert\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:15 crc kubenswrapper[4799]: I1124 06:58:15.995087 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9jhx\" (UniqueName: \"kubernetes.io/projected/64c312a9-19c5-4feb-85e5-95c80da1e647-kube-api-access-t9jhx\") pod \"ovnkube-node-fwwmx\" (UID: \"64c312a9-19c5-4feb-85e5-95c80da1e647\") " pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.030410 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.071070 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" event={"ID":"64c312a9-19c5-4feb-85e5-95c80da1e647","Type":"ContainerStarted","Data":"895b06648adc0a97b4b78df9585b33861899463ef25fee15f32e8619adc7bf16"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.074549 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovnkube-controller/3.log" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.076615 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovn-acl-logging/0.log" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077140 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b7nd7_5c34c957-e45b-4c65-8d6a-60e0a45b2f25/ovn-controller/0.log" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077503 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2" exitCode=0 Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077534 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54" exitCode=0 Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077544 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95" exitCode=0 Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077554 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1" exitCode=0 Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077563 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a" exitCode=0 Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077572 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74" exitCode=0 Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077580 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a" exitCode=143 Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077588 4799 generic.go:334] "Generic (PLEG): container finished" podID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" containerID="4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a" exitCode=143 Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077639 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077677 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077692 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077705 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077741 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077754 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077768 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077781 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077789 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077796 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077803 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077811 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077820 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077828 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077835 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077866 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077878 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077887 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077896 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077903 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077911 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077918 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077925 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077932 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077940 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077947 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077956 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077968 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077979 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077986 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.077994 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078001 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078009 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078016 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078024 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078032 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078040 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078050 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" event={"ID":"5c34c957-e45b-4c65-8d6a-60e0a45b2f25","Type":"ContainerDied","Data":"ed853cc5cdad54443adb0e29e0a8965d4896c34402e739666d0b1d156d04aa85"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078061 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078070 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078077 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078085 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078093 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078101 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078107 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078115 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078122 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078130 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078150 4799 scope.go:117] "RemoveContainer" containerID="f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.078343 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b7nd7" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.082070 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvd7f_7980021c-dde3-4c14-a7b6-fbcc947a183d/kube-multus/2.log" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.083670 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvd7f_7980021c-dde3-4c14-a7b6-fbcc947a183d/kube-multus/1.log" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.083726 4799 generic.go:334] "Generic (PLEG): container finished" podID="7980021c-dde3-4c14-a7b6-fbcc947a183d" containerID="726b153e9aa2f296175c87a8e127b3e7ae73ed29881843f9bb9b1220f6b95d95" exitCode=2 Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.083769 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvd7f" event={"ID":"7980021c-dde3-4c14-a7b6-fbcc947a183d","Type":"ContainerDied","Data":"726b153e9aa2f296175c87a8e127b3e7ae73ed29881843f9bb9b1220f6b95d95"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.083807 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365"} Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.084371 4799 scope.go:117] "RemoveContainer" containerID="726b153e9aa2f296175c87a8e127b3e7ae73ed29881843f9bb9b1220f6b95d95" Nov 24 06:58:16 crc kubenswrapper[4799]: E1124 06:58:16.084613 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-zvd7f_openshift-multus(7980021c-dde3-4c14-a7b6-fbcc947a183d)\"" pod="openshift-multus/multus-zvd7f" podUID="7980021c-dde3-4c14-a7b6-fbcc947a183d" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.105084 4799 scope.go:117] "RemoveContainer" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.133476 4799 scope.go:117] "RemoveContainer" containerID="28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.143734 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b7nd7"] Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.148024 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b7nd7"] Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.156794 4799 scope.go:117] "RemoveContainer" containerID="85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.173089 4799 scope.go:117] "RemoveContainer" containerID="27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.190204 4799 scope.go:117] "RemoveContainer" containerID="771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.211411 4799 scope.go:117] "RemoveContainer" containerID="ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.265873 4799 scope.go:117] "RemoveContainer" containerID="90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.278605 4799 scope.go:117] "RemoveContainer" containerID="4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.301810 4799 scope.go:117] "RemoveContainer" containerID="bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.320613 4799 scope.go:117] "RemoveContainer" containerID="f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2" Nov 24 06:58:16 crc kubenswrapper[4799]: E1124 06:58:16.321400 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2\": container with ID starting with f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2 not found: ID does not exist" containerID="f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.321463 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2"} err="failed to get container status \"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2\": rpc error: code = NotFound desc = could not find container \"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2\": container with ID starting with f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.321498 4799 scope.go:117] "RemoveContainer" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:58:16 crc kubenswrapper[4799]: E1124 06:58:16.322042 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\": container with ID starting with 729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681 not found: ID does not exist" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.322091 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681"} err="failed to get container status \"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\": rpc error: code = NotFound desc = could not find container \"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\": container with ID starting with 729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.322130 4799 scope.go:117] "RemoveContainer" containerID="28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54" Nov 24 06:58:16 crc kubenswrapper[4799]: E1124 06:58:16.322440 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\": container with ID starting with 28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54 not found: ID does not exist" containerID="28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.322471 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54"} err="failed to get container status \"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\": rpc error: code = NotFound desc = could not find container \"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\": container with ID starting with 28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.322490 4799 scope.go:117] "RemoveContainer" containerID="85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95" Nov 24 06:58:16 crc kubenswrapper[4799]: E1124 06:58:16.322914 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\": container with ID starting with 85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95 not found: ID does not exist" containerID="85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.322968 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95"} err="failed to get container status \"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\": rpc error: code = NotFound desc = could not find container \"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\": container with ID starting with 85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.323010 4799 scope.go:117] "RemoveContainer" containerID="27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1" Nov 24 06:58:16 crc kubenswrapper[4799]: E1124 06:58:16.323376 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\": container with ID starting with 27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1 not found: ID does not exist" containerID="27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.323412 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1"} err="failed to get container status \"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\": rpc error: code = NotFound desc = could not find container \"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\": container with ID starting with 27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.323433 4799 scope.go:117] "RemoveContainer" containerID="771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a" Nov 24 06:58:16 crc kubenswrapper[4799]: E1124 06:58:16.323991 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\": container with ID starting with 771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a not found: ID does not exist" containerID="771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.324023 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a"} err="failed to get container status \"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\": rpc error: code = NotFound desc = could not find container \"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\": container with ID starting with 771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.324047 4799 scope.go:117] "RemoveContainer" containerID="ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74" Nov 24 06:58:16 crc kubenswrapper[4799]: E1124 06:58:16.324401 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\": container with ID starting with ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74 not found: ID does not exist" containerID="ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.324472 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74"} err="failed to get container status \"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\": rpc error: code = NotFound desc = could not find container \"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\": container with ID starting with ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.324525 4799 scope.go:117] "RemoveContainer" containerID="90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a" Nov 24 06:58:16 crc kubenswrapper[4799]: E1124 06:58:16.324952 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\": container with ID starting with 90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a not found: ID does not exist" containerID="90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.324985 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a"} err="failed to get container status \"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\": rpc error: code = NotFound desc = could not find container \"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\": container with ID starting with 90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.325019 4799 scope.go:117] "RemoveContainer" containerID="4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a" Nov 24 06:58:16 crc kubenswrapper[4799]: E1124 06:58:16.325560 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\": container with ID starting with 4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a not found: ID does not exist" containerID="4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.325594 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a"} err="failed to get container status \"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\": rpc error: code = NotFound desc = could not find container \"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\": container with ID starting with 4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.325616 4799 scope.go:117] "RemoveContainer" containerID="bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c" Nov 24 06:58:16 crc kubenswrapper[4799]: E1124 06:58:16.326356 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\": container with ID starting with bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c not found: ID does not exist" containerID="bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.326387 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c"} err="failed to get container status \"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\": rpc error: code = NotFound desc = could not find container \"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\": container with ID starting with bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.326404 4799 scope.go:117] "RemoveContainer" containerID="f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.326778 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2"} err="failed to get container status \"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2\": rpc error: code = NotFound desc = could not find container \"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2\": container with ID starting with f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.326802 4799 scope.go:117] "RemoveContainer" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.327161 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681"} err="failed to get container status \"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\": rpc error: code = NotFound desc = could not find container \"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\": container with ID starting with 729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.327190 4799 scope.go:117] "RemoveContainer" containerID="28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.327513 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54"} err="failed to get container status \"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\": rpc error: code = NotFound desc = could not find container \"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\": container with ID starting with 28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.327538 4799 scope.go:117] "RemoveContainer" containerID="85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.327893 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95"} err="failed to get container status \"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\": rpc error: code = NotFound desc = could not find container \"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\": container with ID starting with 85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.327914 4799 scope.go:117] "RemoveContainer" containerID="27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.328244 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1"} err="failed to get container status \"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\": rpc error: code = NotFound desc = could not find container \"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\": container with ID starting with 27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.328268 4799 scope.go:117] "RemoveContainer" containerID="771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.328577 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a"} err="failed to get container status \"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\": rpc error: code = NotFound desc = could not find container \"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\": container with ID starting with 771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.328598 4799 scope.go:117] "RemoveContainer" containerID="ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.328937 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74"} err="failed to get container status \"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\": rpc error: code = NotFound desc = could not find container \"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\": container with ID starting with ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.328959 4799 scope.go:117] "RemoveContainer" containerID="90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.329256 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a"} err="failed to get container status \"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\": rpc error: code = NotFound desc = could not find container \"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\": container with ID starting with 90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.329288 4799 scope.go:117] "RemoveContainer" containerID="4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.329569 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a"} err="failed to get container status \"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\": rpc error: code = NotFound desc = could not find container \"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\": container with ID starting with 4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.329655 4799 scope.go:117] "RemoveContainer" containerID="bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.330193 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c"} err="failed to get container status \"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\": rpc error: code = NotFound desc = could not find container \"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\": container with ID starting with bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.330217 4799 scope.go:117] "RemoveContainer" containerID="f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.330825 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2"} err="failed to get container status \"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2\": rpc error: code = NotFound desc = could not find container \"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2\": container with ID starting with f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.330868 4799 scope.go:117] "RemoveContainer" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.331289 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681"} err="failed to get container status \"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\": rpc error: code = NotFound desc = could not find container \"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\": container with ID starting with 729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.331330 4799 scope.go:117] "RemoveContainer" containerID="28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.331795 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54"} err="failed to get container status \"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\": rpc error: code = NotFound desc = could not find container \"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\": container with ID starting with 28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.331818 4799 scope.go:117] "RemoveContainer" containerID="85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.332741 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95"} err="failed to get container status \"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\": rpc error: code = NotFound desc = could not find container \"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\": container with ID starting with 85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.332767 4799 scope.go:117] "RemoveContainer" containerID="27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.333122 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1"} err="failed to get container status \"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\": rpc error: code = NotFound desc = could not find container \"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\": container with ID starting with 27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.333144 4799 scope.go:117] "RemoveContainer" containerID="771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.333772 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a"} err="failed to get container status \"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\": rpc error: code = NotFound desc = could not find container \"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\": container with ID starting with 771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.333800 4799 scope.go:117] "RemoveContainer" containerID="ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.334135 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74"} err="failed to get container status \"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\": rpc error: code = NotFound desc = could not find container \"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\": container with ID starting with ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.334170 4799 scope.go:117] "RemoveContainer" containerID="90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.334652 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a"} err="failed to get container status \"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\": rpc error: code = NotFound desc = could not find container \"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\": container with ID starting with 90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.334680 4799 scope.go:117] "RemoveContainer" containerID="4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.334988 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a"} err="failed to get container status \"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\": rpc error: code = NotFound desc = could not find container \"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\": container with ID starting with 4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.335011 4799 scope.go:117] "RemoveContainer" containerID="bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.335600 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c"} err="failed to get container status \"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\": rpc error: code = NotFound desc = could not find container \"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\": container with ID starting with bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.335627 4799 scope.go:117] "RemoveContainer" containerID="f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.336535 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2"} err="failed to get container status \"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2\": rpc error: code = NotFound desc = could not find container \"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2\": container with ID starting with f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.336560 4799 scope.go:117] "RemoveContainer" containerID="729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.336992 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681"} err="failed to get container status \"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\": rpc error: code = NotFound desc = could not find container \"729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681\": container with ID starting with 729a4f1251c556ccc2596b836267b6d4b4cceeada59ae66b6457df82834ff681 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.337018 4799 scope.go:117] "RemoveContainer" containerID="28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.337551 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54"} err="failed to get container status \"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\": rpc error: code = NotFound desc = could not find container \"28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54\": container with ID starting with 28b712ff01fc9433c1c5b0ae5a48a179a9bc180575e60f9a9af1417dd8541b54 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.337574 4799 scope.go:117] "RemoveContainer" containerID="85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.339825 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95"} err="failed to get container status \"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\": rpc error: code = NotFound desc = could not find container \"85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95\": container with ID starting with 85a4fa565d00d6d843d293ad4b950e50e4334e5902736c6d02f2d54f8a89de95 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.339872 4799 scope.go:117] "RemoveContainer" containerID="27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.340407 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1"} err="failed to get container status \"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\": rpc error: code = NotFound desc = could not find container \"27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1\": container with ID starting with 27f118e6eba06912ce50108154ff3472084b18661e39eb6a2a4d560cdd7b5fc1 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.340439 4799 scope.go:117] "RemoveContainer" containerID="771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.341173 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a"} err="failed to get container status \"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\": rpc error: code = NotFound desc = could not find container \"771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a\": container with ID starting with 771ccda264ff8ee28154094fcb6c5a3827c63c6de1659fca7e7fd3e7d327e28a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.341198 4799 scope.go:117] "RemoveContainer" containerID="ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.341522 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74"} err="failed to get container status \"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\": rpc error: code = NotFound desc = could not find container \"ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74\": container with ID starting with ee57f33ad5e4b1725122541a7517ad5bf158d0d2cdabb91657a1bca2662aad74 not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.341556 4799 scope.go:117] "RemoveContainer" containerID="90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.341939 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a"} err="failed to get container status \"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\": rpc error: code = NotFound desc = could not find container \"90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a\": container with ID starting with 90f08f2e748c1a6f96ccaf9d512749a9a1c5521944a83afab2b1a772bbba367a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.341975 4799 scope.go:117] "RemoveContainer" containerID="4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.342272 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a"} err="failed to get container status \"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\": rpc error: code = NotFound desc = could not find container \"4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a\": container with ID starting with 4ef05b3ce7054a3653fc8ce0e0dd1aede3317ad46d4d367ab08ff2297467167a not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.342299 4799 scope.go:117] "RemoveContainer" containerID="bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.342763 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c"} err="failed to get container status \"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\": rpc error: code = NotFound desc = could not find container \"bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c\": container with ID starting with bbebc1620deecd6e8889925822194e3153c99e21e0d9acebd8ae18f3ecacee9c not found: ID does not exist" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.342803 4799 scope.go:117] "RemoveContainer" containerID="f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2" Nov 24 06:58:16 crc kubenswrapper[4799]: I1124 06:58:16.343534 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2"} err="failed to get container status \"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2\": rpc error: code = NotFound desc = could not find container \"f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2\": container with ID starting with f32dd2acde0f9d19d6b5758b695a0541ce691a6900c0e67b67de740edc6f0bf2 not found: ID does not exist" Nov 24 06:58:17 crc kubenswrapper[4799]: I1124 06:58:17.096254 4799 generic.go:334] "Generic (PLEG): container finished" podID="64c312a9-19c5-4feb-85e5-95c80da1e647" containerID="69b69221f14da7dd7435c1e71480d20eb393737323093d64cf23184ee30a51a6" exitCode=0 Nov 24 06:58:17 crc kubenswrapper[4799]: I1124 06:58:17.096411 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" event={"ID":"64c312a9-19c5-4feb-85e5-95c80da1e647","Type":"ContainerDied","Data":"69b69221f14da7dd7435c1e71480d20eb393737323093d64cf23184ee30a51a6"} Nov 24 06:58:17 crc kubenswrapper[4799]: I1124 06:58:17.638390 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c34c957-e45b-4c65-8d6a-60e0a45b2f25" path="/var/lib/kubelet/pods/5c34c957-e45b-4c65-8d6a-60e0a45b2f25/volumes" Nov 24 06:58:18 crc kubenswrapper[4799]: I1124 06:58:18.106234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" event={"ID":"64c312a9-19c5-4feb-85e5-95c80da1e647","Type":"ContainerStarted","Data":"a32982f567e74aab736d2009272e26a06e300fe2e0fe144108e5155a20bb2ad4"} Nov 24 06:58:18 crc kubenswrapper[4799]: I1124 06:58:18.106734 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" event={"ID":"64c312a9-19c5-4feb-85e5-95c80da1e647","Type":"ContainerStarted","Data":"68e2656bf8834fad6689dfc832a313905c86da0ee1ca56830f23b533fc5bf43e"} Nov 24 06:58:18 crc kubenswrapper[4799]: I1124 06:58:18.106751 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" event={"ID":"64c312a9-19c5-4feb-85e5-95c80da1e647","Type":"ContainerStarted","Data":"a8a0f7ef16a2644dafbe4193358d515f102198dc94b12a2ff7423e0966647f34"} Nov 24 06:58:18 crc kubenswrapper[4799]: I1124 06:58:18.106766 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" event={"ID":"64c312a9-19c5-4feb-85e5-95c80da1e647","Type":"ContainerStarted","Data":"a8682555c5c8147282b465766d848f9fd06813f624c0f215f3fab0c38c268a56"} Nov 24 06:58:18 crc kubenswrapper[4799]: I1124 06:58:18.106779 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" event={"ID":"64c312a9-19c5-4feb-85e5-95c80da1e647","Type":"ContainerStarted","Data":"403622c3bebecf41752495dc94d57713cda07f244ad3f51bc250f72677ae7ad5"} Nov 24 06:58:18 crc kubenswrapper[4799]: I1124 06:58:18.106791 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" event={"ID":"64c312a9-19c5-4feb-85e5-95c80da1e647","Type":"ContainerStarted","Data":"dd4f242fad98c61eef00b5723f5fbfbf58c3ff63afb668968484156b43c41bef"} Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.691968 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-pvnqk"] Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.694703 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.697768 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.698168 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.698243 4799 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-2smct" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.698186 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.849765 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/65589a34-a1fb-49cb-ba7b-57da57096604-crc-storage\") pod \"crc-storage-crc-pvnqk\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.849921 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qgd4\" (UniqueName: \"kubernetes.io/projected/65589a34-a1fb-49cb-ba7b-57da57096604-kube-api-access-2qgd4\") pod \"crc-storage-crc-pvnqk\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.850064 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/65589a34-a1fb-49cb-ba7b-57da57096604-node-mnt\") pod \"crc-storage-crc-pvnqk\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.951758 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qgd4\" (UniqueName: \"kubernetes.io/projected/65589a34-a1fb-49cb-ba7b-57da57096604-kube-api-access-2qgd4\") pod \"crc-storage-crc-pvnqk\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.952120 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/65589a34-a1fb-49cb-ba7b-57da57096604-node-mnt\") pod \"crc-storage-crc-pvnqk\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.952293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/65589a34-a1fb-49cb-ba7b-57da57096604-crc-storage\") pod \"crc-storage-crc-pvnqk\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.952674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/65589a34-a1fb-49cb-ba7b-57da57096604-node-mnt\") pod \"crc-storage-crc-pvnqk\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.953448 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/65589a34-a1fb-49cb-ba7b-57da57096604-crc-storage\") pod \"crc-storage-crc-pvnqk\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:20 crc kubenswrapper[4799]: I1124 06:58:20.989142 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qgd4\" (UniqueName: \"kubernetes.io/projected/65589a34-a1fb-49cb-ba7b-57da57096604-kube-api-access-2qgd4\") pod \"crc-storage-crc-pvnqk\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:21 crc kubenswrapper[4799]: I1124 06:58:21.019409 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:21 crc kubenswrapper[4799]: E1124 06:58:21.057971 4799 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(e5201e04dae3332850d9d7d3189688a14f380078f0b10ec979fe711df4bae263): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 06:58:21 crc kubenswrapper[4799]: E1124 06:58:21.058106 4799 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(e5201e04dae3332850d9d7d3189688a14f380078f0b10ec979fe711df4bae263): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:21 crc kubenswrapper[4799]: E1124 06:58:21.058147 4799 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(e5201e04dae3332850d9d7d3189688a14f380078f0b10ec979fe711df4bae263): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:21 crc kubenswrapper[4799]: E1124 06:58:21.058506 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-pvnqk_crc-storage(65589a34-a1fb-49cb-ba7b-57da57096604)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-pvnqk_crc-storage(65589a34-a1fb-49cb-ba7b-57da57096604)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(e5201e04dae3332850d9d7d3189688a14f380078f0b10ec979fe711df4bae263): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-pvnqk" podUID="65589a34-a1fb-49cb-ba7b-57da57096604" Nov 24 06:58:21 crc kubenswrapper[4799]: I1124 06:58:21.133077 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" event={"ID":"64c312a9-19c5-4feb-85e5-95c80da1e647","Type":"ContainerStarted","Data":"161ffff961e172320f627cc36433264de5ebfa7f66c600f1f5bc58c5f8fa6357"} Nov 24 06:58:23 crc kubenswrapper[4799]: I1124 06:58:23.116940 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-pvnqk"] Nov 24 06:58:23 crc kubenswrapper[4799]: I1124 06:58:23.117119 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:23 crc kubenswrapper[4799]: I1124 06:58:23.117771 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:23 crc kubenswrapper[4799]: E1124 06:58:23.149098 4799 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(233318d53fd46fdfdaa80256a3b279256601f24199794dc9e9e35499da14b51b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 06:58:23 crc kubenswrapper[4799]: E1124 06:58:23.149177 4799 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(233318d53fd46fdfdaa80256a3b279256601f24199794dc9e9e35499da14b51b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:23 crc kubenswrapper[4799]: E1124 06:58:23.149217 4799 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(233318d53fd46fdfdaa80256a3b279256601f24199794dc9e9e35499da14b51b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:23 crc kubenswrapper[4799]: E1124 06:58:23.149312 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-pvnqk_crc-storage(65589a34-a1fb-49cb-ba7b-57da57096604)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-pvnqk_crc-storage(65589a34-a1fb-49cb-ba7b-57da57096604)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(233318d53fd46fdfdaa80256a3b279256601f24199794dc9e9e35499da14b51b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-pvnqk" podUID="65589a34-a1fb-49cb-ba7b-57da57096604" Nov 24 06:58:23 crc kubenswrapper[4799]: I1124 06:58:23.151778 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" event={"ID":"64c312a9-19c5-4feb-85e5-95c80da1e647","Type":"ContainerStarted","Data":"11a7c10276b54ce202bf95d5a2fe31e71ad480172664c84d7f831f53e6b1243b"} Nov 24 06:58:23 crc kubenswrapper[4799]: I1124 06:58:23.152428 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:23 crc kubenswrapper[4799]: I1124 06:58:23.152470 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:23 crc kubenswrapper[4799]: I1124 06:58:23.188276 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:23 crc kubenswrapper[4799]: I1124 06:58:23.208565 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" podStartSLOduration=8.20853718 podStartE2EDuration="8.20853718s" podCreationTimestamp="2025-11-24 06:58:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:58:23.198947413 +0000 UTC m=+648.854929907" watchObservedRunningTime="2025-11-24 06:58:23.20853718 +0000 UTC m=+648.864519694" Nov 24 06:58:24 crc kubenswrapper[4799]: I1124 06:58:24.160421 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:24 crc kubenswrapper[4799]: I1124 06:58:24.196249 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:27 crc kubenswrapper[4799]: I1124 06:58:27.628712 4799 scope.go:117] "RemoveContainer" containerID="726b153e9aa2f296175c87a8e127b3e7ae73ed29881843f9bb9b1220f6b95d95" Nov 24 06:58:27 crc kubenswrapper[4799]: E1124 06:58:27.628986 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-zvd7f_openshift-multus(7980021c-dde3-4c14-a7b6-fbcc947a183d)\"" pod="openshift-multus/multus-zvd7f" podUID="7980021c-dde3-4c14-a7b6-fbcc947a183d" Nov 24 06:58:35 crc kubenswrapper[4799]: I1124 06:58:35.742510 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:35 crc kubenswrapper[4799]: I1124 06:58:35.743328 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:35 crc kubenswrapper[4799]: E1124 06:58:35.785332 4799 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(87544db528217bfd8495f975e59cad95e84fff0901ebbcaeb251184cb8637afe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 06:58:35 crc kubenswrapper[4799]: E1124 06:58:35.785494 4799 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(87544db528217bfd8495f975e59cad95e84fff0901ebbcaeb251184cb8637afe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:35 crc kubenswrapper[4799]: E1124 06:58:35.785551 4799 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(87544db528217bfd8495f975e59cad95e84fff0901ebbcaeb251184cb8637afe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:35 crc kubenswrapper[4799]: E1124 06:58:35.785675 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-pvnqk_crc-storage(65589a34-a1fb-49cb-ba7b-57da57096604)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-pvnqk_crc-storage(65589a34-a1fb-49cb-ba7b-57da57096604)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pvnqk_crc-storage_65589a34-a1fb-49cb-ba7b-57da57096604_0(87544db528217bfd8495f975e59cad95e84fff0901ebbcaeb251184cb8637afe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-pvnqk" podUID="65589a34-a1fb-49cb-ba7b-57da57096604" Nov 24 06:58:36 crc kubenswrapper[4799]: I1124 06:58:36.257758 4799 scope.go:117] "RemoveContainer" containerID="ae20e1630b8a28aa4daa83fd5e4380ecf88e0af733819883b134e6094c178365" Nov 24 06:58:37 crc kubenswrapper[4799]: I1124 06:58:37.270743 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvd7f_7980021c-dde3-4c14-a7b6-fbcc947a183d/kube-multus/2.log" Nov 24 06:58:42 crc kubenswrapper[4799]: I1124 06:58:42.629323 4799 scope.go:117] "RemoveContainer" containerID="726b153e9aa2f296175c87a8e127b3e7ae73ed29881843f9bb9b1220f6b95d95" Nov 24 06:58:43 crc kubenswrapper[4799]: I1124 06:58:43.323633 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvd7f_7980021c-dde3-4c14-a7b6-fbcc947a183d/kube-multus/2.log" Nov 24 06:58:43 crc kubenswrapper[4799]: I1124 06:58:43.324379 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvd7f" event={"ID":"7980021c-dde3-4c14-a7b6-fbcc947a183d","Type":"ContainerStarted","Data":"36034c4db42b81526922b8ffd7c74efd68c440c65d2aadfad0c6ecd095332925"} Nov 24 06:58:46 crc kubenswrapper[4799]: I1124 06:58:46.066938 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fwwmx" Nov 24 06:58:50 crc kubenswrapper[4799]: I1124 06:58:50.627590 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:50 crc kubenswrapper[4799]: I1124 06:58:50.629284 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:50 crc kubenswrapper[4799]: I1124 06:58:50.856518 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-pvnqk"] Nov 24 06:58:50 crc kubenswrapper[4799]: I1124 06:58:50.868259 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 06:58:51 crc kubenswrapper[4799]: I1124 06:58:51.377440 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-pvnqk" event={"ID":"65589a34-a1fb-49cb-ba7b-57da57096604","Type":"ContainerStarted","Data":"fb6dffead31405905b35c57c31c70205a1f8bd4835acf97032922d0fb30c9eed"} Nov 24 06:58:52 crc kubenswrapper[4799]: I1124 06:58:52.388886 4799 generic.go:334] "Generic (PLEG): container finished" podID="65589a34-a1fb-49cb-ba7b-57da57096604" containerID="f3b7bfe56f907040c8123c075eae2d219a2ae33843b9e40f05d41179e21aa19c" exitCode=0 Nov 24 06:58:52 crc kubenswrapper[4799]: I1124 06:58:52.389031 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-pvnqk" event={"ID":"65589a34-a1fb-49cb-ba7b-57da57096604","Type":"ContainerDied","Data":"f3b7bfe56f907040c8123c075eae2d219a2ae33843b9e40f05d41179e21aa19c"} Nov 24 06:58:53 crc kubenswrapper[4799]: I1124 06:58:53.742698 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:58:53 crc kubenswrapper[4799]: I1124 06:58:53.757472 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qgd4\" (UniqueName: \"kubernetes.io/projected/65589a34-a1fb-49cb-ba7b-57da57096604-kube-api-access-2qgd4\") pod \"65589a34-a1fb-49cb-ba7b-57da57096604\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " Nov 24 06:58:53 crc kubenswrapper[4799]: I1124 06:58:53.757679 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/65589a34-a1fb-49cb-ba7b-57da57096604-crc-storage\") pod \"65589a34-a1fb-49cb-ba7b-57da57096604\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " Nov 24 06:58:53 crc kubenswrapper[4799]: I1124 06:58:53.757740 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/65589a34-a1fb-49cb-ba7b-57da57096604-node-mnt\") pod \"65589a34-a1fb-49cb-ba7b-57da57096604\" (UID: \"65589a34-a1fb-49cb-ba7b-57da57096604\") " Nov 24 06:58:53 crc kubenswrapper[4799]: I1124 06:58:53.758261 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65589a34-a1fb-49cb-ba7b-57da57096604-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "65589a34-a1fb-49cb-ba7b-57da57096604" (UID: "65589a34-a1fb-49cb-ba7b-57da57096604"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 06:58:53 crc kubenswrapper[4799]: I1124 06:58:53.769121 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65589a34-a1fb-49cb-ba7b-57da57096604-kube-api-access-2qgd4" (OuterVolumeSpecName: "kube-api-access-2qgd4") pod "65589a34-a1fb-49cb-ba7b-57da57096604" (UID: "65589a34-a1fb-49cb-ba7b-57da57096604"). InnerVolumeSpecName "kube-api-access-2qgd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:58:53 crc kubenswrapper[4799]: I1124 06:58:53.776202 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65589a34-a1fb-49cb-ba7b-57da57096604-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "65589a34-a1fb-49cb-ba7b-57da57096604" (UID: "65589a34-a1fb-49cb-ba7b-57da57096604"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:58:53 crc kubenswrapper[4799]: I1124 06:58:53.859944 4799 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/65589a34-a1fb-49cb-ba7b-57da57096604-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:53 crc kubenswrapper[4799]: I1124 06:58:53.859995 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qgd4\" (UniqueName: \"kubernetes.io/projected/65589a34-a1fb-49cb-ba7b-57da57096604-kube-api-access-2qgd4\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:53 crc kubenswrapper[4799]: I1124 06:58:53.860008 4799 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/65589a34-a1fb-49cb-ba7b-57da57096604-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 24 06:58:54 crc kubenswrapper[4799]: I1124 06:58:54.409234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-pvnqk" event={"ID":"65589a34-a1fb-49cb-ba7b-57da57096604","Type":"ContainerDied","Data":"fb6dffead31405905b35c57c31c70205a1f8bd4835acf97032922d0fb30c9eed"} Nov 24 06:58:54 crc kubenswrapper[4799]: I1124 06:58:54.409311 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb6dffead31405905b35c57c31c70205a1f8bd4835acf97032922d0fb30c9eed" Nov 24 06:58:54 crc kubenswrapper[4799]: I1124 06:58:54.409340 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pvnqk" Nov 24 06:59:00 crc kubenswrapper[4799]: I1124 06:59:00.970880 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr"] Nov 24 06:59:00 crc kubenswrapper[4799]: E1124 06:59:00.972088 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65589a34-a1fb-49cb-ba7b-57da57096604" containerName="storage" Nov 24 06:59:00 crc kubenswrapper[4799]: I1124 06:59:00.972110 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="65589a34-a1fb-49cb-ba7b-57da57096604" containerName="storage" Nov 24 06:59:00 crc kubenswrapper[4799]: I1124 06:59:00.972231 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="65589a34-a1fb-49cb-ba7b-57da57096604" containerName="storage" Nov 24 06:59:00 crc kubenswrapper[4799]: I1124 06:59:00.973309 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:00 crc kubenswrapper[4799]: I1124 06:59:00.977005 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 06:59:00 crc kubenswrapper[4799]: I1124 06:59:00.984339 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr"] Nov 24 06:59:01 crc kubenswrapper[4799]: I1124 06:59:01.167397 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:01 crc kubenswrapper[4799]: I1124 06:59:01.167539 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:01 crc kubenswrapper[4799]: I1124 06:59:01.167585 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26lzs\" (UniqueName: \"kubernetes.io/projected/e055d22a-1e71-499f-a898-fdcff0b33a35-kube-api-access-26lzs\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:01 crc kubenswrapper[4799]: I1124 06:59:01.269287 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26lzs\" (UniqueName: \"kubernetes.io/projected/e055d22a-1e71-499f-a898-fdcff0b33a35-kube-api-access-26lzs\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:01 crc kubenswrapper[4799]: I1124 06:59:01.269406 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:01 crc kubenswrapper[4799]: I1124 06:59:01.269583 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:01 crc kubenswrapper[4799]: I1124 06:59:01.270691 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:01 crc kubenswrapper[4799]: I1124 06:59:01.270710 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:01 crc kubenswrapper[4799]: I1124 06:59:01.298948 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26lzs\" (UniqueName: \"kubernetes.io/projected/e055d22a-1e71-499f-a898-fdcff0b33a35-kube-api-access-26lzs\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:01 crc kubenswrapper[4799]: I1124 06:59:01.590111 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:01 crc kubenswrapper[4799]: I1124 06:59:01.870976 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr"] Nov 24 06:59:02 crc kubenswrapper[4799]: I1124 06:59:02.469829 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" event={"ID":"e055d22a-1e71-499f-a898-fdcff0b33a35","Type":"ContainerStarted","Data":"9ef7f5df13e377cdac01a2f29cc4b4a6c516e9889b4e9014d87b3a2143abf5ad"} Nov 24 06:59:02 crc kubenswrapper[4799]: I1124 06:59:02.469958 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" event={"ID":"e055d22a-1e71-499f-a898-fdcff0b33a35","Type":"ContainerStarted","Data":"5951e7c63b78dacb2cc7f577898c618bc9494cef66111efe33381d51b71fa6f7"} Nov 24 06:59:03 crc kubenswrapper[4799]: I1124 06:59:03.477596 4799 generic.go:334] "Generic (PLEG): container finished" podID="e055d22a-1e71-499f-a898-fdcff0b33a35" containerID="9ef7f5df13e377cdac01a2f29cc4b4a6c516e9889b4e9014d87b3a2143abf5ad" exitCode=0 Nov 24 06:59:03 crc kubenswrapper[4799]: I1124 06:59:03.477713 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" event={"ID":"e055d22a-1e71-499f-a898-fdcff0b33a35","Type":"ContainerDied","Data":"9ef7f5df13e377cdac01a2f29cc4b4a6c516e9889b4e9014d87b3a2143abf5ad"} Nov 24 06:59:04 crc kubenswrapper[4799]: E1124 06:59:04.802574 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode055d22a_1e71_499f_a898_fdcff0b33a35.slice/crio-07d4dbd79529d6e8a50647502040a89ac58f85688a504daabb1ed0ce9e43de74.scope\": RecentStats: unable to find data in memory cache]" Nov 24 06:59:05 crc kubenswrapper[4799]: I1124 06:59:05.494465 4799 generic.go:334] "Generic (PLEG): container finished" podID="e055d22a-1e71-499f-a898-fdcff0b33a35" containerID="07d4dbd79529d6e8a50647502040a89ac58f85688a504daabb1ed0ce9e43de74" exitCode=0 Nov 24 06:59:05 crc kubenswrapper[4799]: I1124 06:59:05.494543 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" event={"ID":"e055d22a-1e71-499f-a898-fdcff0b33a35","Type":"ContainerDied","Data":"07d4dbd79529d6e8a50647502040a89ac58f85688a504daabb1ed0ce9e43de74"} Nov 24 06:59:06 crc kubenswrapper[4799]: I1124 06:59:06.506610 4799 generic.go:334] "Generic (PLEG): container finished" podID="e055d22a-1e71-499f-a898-fdcff0b33a35" containerID="4774f59ab9acbf3b0513ddc40351167cbb59b52fa6937e1dae0ac4e5c9d01826" exitCode=0 Nov 24 06:59:06 crc kubenswrapper[4799]: I1124 06:59:06.506672 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" event={"ID":"e055d22a-1e71-499f-a898-fdcff0b33a35","Type":"ContainerDied","Data":"4774f59ab9acbf3b0513ddc40351167cbb59b52fa6937e1dae0ac4e5c9d01826"} Nov 24 06:59:07 crc kubenswrapper[4799]: I1124 06:59:07.824016 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:07 crc kubenswrapper[4799]: I1124 06:59:07.972738 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26lzs\" (UniqueName: \"kubernetes.io/projected/e055d22a-1e71-499f-a898-fdcff0b33a35-kube-api-access-26lzs\") pod \"e055d22a-1e71-499f-a898-fdcff0b33a35\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " Nov 24 06:59:07 crc kubenswrapper[4799]: I1124 06:59:07.972875 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-bundle\") pod \"e055d22a-1e71-499f-a898-fdcff0b33a35\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " Nov 24 06:59:07 crc kubenswrapper[4799]: I1124 06:59:07.972918 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-util\") pod \"e055d22a-1e71-499f-a898-fdcff0b33a35\" (UID: \"e055d22a-1e71-499f-a898-fdcff0b33a35\") " Nov 24 06:59:07 crc kubenswrapper[4799]: I1124 06:59:07.973832 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-bundle" (OuterVolumeSpecName: "bundle") pod "e055d22a-1e71-499f-a898-fdcff0b33a35" (UID: "e055d22a-1e71-499f-a898-fdcff0b33a35"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:07 crc kubenswrapper[4799]: I1124 06:59:07.974657 4799 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:07 crc kubenswrapper[4799]: I1124 06:59:07.982616 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e055d22a-1e71-499f-a898-fdcff0b33a35-kube-api-access-26lzs" (OuterVolumeSpecName: "kube-api-access-26lzs") pod "e055d22a-1e71-499f-a898-fdcff0b33a35" (UID: "e055d22a-1e71-499f-a898-fdcff0b33a35"). InnerVolumeSpecName "kube-api-access-26lzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:59:07 crc kubenswrapper[4799]: I1124 06:59:07.984296 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-util" (OuterVolumeSpecName: "util") pod "e055d22a-1e71-499f-a898-fdcff0b33a35" (UID: "e055d22a-1e71-499f-a898-fdcff0b33a35"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:08 crc kubenswrapper[4799]: I1124 06:59:08.075534 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26lzs\" (UniqueName: \"kubernetes.io/projected/e055d22a-1e71-499f-a898-fdcff0b33a35-kube-api-access-26lzs\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:08 crc kubenswrapper[4799]: I1124 06:59:08.075600 4799 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e055d22a-1e71-499f-a898-fdcff0b33a35-util\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:08 crc kubenswrapper[4799]: I1124 06:59:08.526407 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" event={"ID":"e055d22a-1e71-499f-a898-fdcff0b33a35","Type":"ContainerDied","Data":"5951e7c63b78dacb2cc7f577898c618bc9494cef66111efe33381d51b71fa6f7"} Nov 24 06:59:08 crc kubenswrapper[4799]: I1124 06:59:08.526498 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5951e7c63b78dacb2cc7f577898c618bc9494cef66111efe33381d51b71fa6f7" Nov 24 06:59:08 crc kubenswrapper[4799]: I1124 06:59:08.526515 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.537339 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-8swlg"] Nov 24 06:59:09 crc kubenswrapper[4799]: E1124 06:59:09.537642 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e055d22a-1e71-499f-a898-fdcff0b33a35" containerName="pull" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.537659 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e055d22a-1e71-499f-a898-fdcff0b33a35" containerName="pull" Nov 24 06:59:09 crc kubenswrapper[4799]: E1124 06:59:09.537682 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e055d22a-1e71-499f-a898-fdcff0b33a35" containerName="util" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.537689 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e055d22a-1e71-499f-a898-fdcff0b33a35" containerName="util" Nov 24 06:59:09 crc kubenswrapper[4799]: E1124 06:59:09.537705 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e055d22a-1e71-499f-a898-fdcff0b33a35" containerName="extract" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.537712 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e055d22a-1e71-499f-a898-fdcff0b33a35" containerName="extract" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.537860 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e055d22a-1e71-499f-a898-fdcff0b33a35" containerName="extract" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.538455 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-8swlg" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.541889 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.541889 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.542159 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-qt79j" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.560444 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-8swlg"] Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.597569 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9khq\" (UniqueName: \"kubernetes.io/projected/d89cee1d-8190-4238-a6ae-f8b629e40cc0-kube-api-access-j9khq\") pod \"nmstate-operator-557fdffb88-8swlg\" (UID: \"d89cee1d-8190-4238-a6ae-f8b629e40cc0\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-8swlg" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.698948 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9khq\" (UniqueName: \"kubernetes.io/projected/d89cee1d-8190-4238-a6ae-f8b629e40cc0-kube-api-access-j9khq\") pod \"nmstate-operator-557fdffb88-8swlg\" (UID: \"d89cee1d-8190-4238-a6ae-f8b629e40cc0\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-8swlg" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.716424 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9khq\" (UniqueName: \"kubernetes.io/projected/d89cee1d-8190-4238-a6ae-f8b629e40cc0-kube-api-access-j9khq\") pod \"nmstate-operator-557fdffb88-8swlg\" (UID: \"d89cee1d-8190-4238-a6ae-f8b629e40cc0\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-8swlg" Nov 24 06:59:09 crc kubenswrapper[4799]: I1124 06:59:09.854617 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-8swlg" Nov 24 06:59:10 crc kubenswrapper[4799]: I1124 06:59:10.033541 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-8swlg"] Nov 24 06:59:10 crc kubenswrapper[4799]: W1124 06:59:10.041831 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd89cee1d_8190_4238_a6ae_f8b629e40cc0.slice/crio-19af16b3b9434fc2cbb2b07eb090d339acdb84611b5d3efe1e7542b9c3a1d391 WatchSource:0}: Error finding container 19af16b3b9434fc2cbb2b07eb090d339acdb84611b5d3efe1e7542b9c3a1d391: Status 404 returned error can't find the container with id 19af16b3b9434fc2cbb2b07eb090d339acdb84611b5d3efe1e7542b9c3a1d391 Nov 24 06:59:10 crc kubenswrapper[4799]: I1124 06:59:10.538237 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-8swlg" event={"ID":"d89cee1d-8190-4238-a6ae-f8b629e40cc0","Type":"ContainerStarted","Data":"19af16b3b9434fc2cbb2b07eb090d339acdb84611b5d3efe1e7542b9c3a1d391"} Nov 24 06:59:12 crc kubenswrapper[4799]: I1124 06:59:12.551806 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-8swlg" event={"ID":"d89cee1d-8190-4238-a6ae-f8b629e40cc0","Type":"ContainerStarted","Data":"643d24f6fcb0576edbc639b58fee1cbc9e7a93df8b3fbaa98abc9d38f9642e6f"} Nov 24 06:59:12 crc kubenswrapper[4799]: I1124 06:59:12.570710 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-8swlg" podStartSLOduration=1.8859706200000002 podStartE2EDuration="3.570684643s" podCreationTimestamp="2025-11-24 06:59:09 +0000 UTC" firstStartedPulling="2025-11-24 06:59:10.045687571 +0000 UTC m=+695.701670045" lastFinishedPulling="2025-11-24 06:59:11.730401594 +0000 UTC m=+697.386384068" observedRunningTime="2025-11-24 06:59:12.566722761 +0000 UTC m=+698.222705235" watchObservedRunningTime="2025-11-24 06:59:12.570684643 +0000 UTC m=+698.226667117" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.570409 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j"] Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.571406 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.575356 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-l5j6c" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.593349 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27"] Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.594316 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.596808 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.613006 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-p9xqt"] Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.614343 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.666213 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j"] Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.673357 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27"] Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.758127 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4f03b272-18b2-423e-9e5d-da5fb982df9c-ovs-socket\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.758206 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/72c88d6c-d53a-47e1-a1e2-4b7292c814b4-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-h7t27\" (UID: \"72c88d6c-d53a-47e1-a1e2-4b7292c814b4\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.758264 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4f03b272-18b2-423e-9e5d-da5fb982df9c-dbus-socket\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.758370 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4f03b272-18b2-423e-9e5d-da5fb982df9c-nmstate-lock\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.758407 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjj2x\" (UniqueName: \"kubernetes.io/projected/72c88d6c-d53a-47e1-a1e2-4b7292c814b4-kube-api-access-fjj2x\") pod \"nmstate-webhook-6b89b748d8-h7t27\" (UID: \"72c88d6c-d53a-47e1-a1e2-4b7292c814b4\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.758439 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48gn8\" (UniqueName: \"kubernetes.io/projected/4f03b272-18b2-423e-9e5d-da5fb982df9c-kube-api-access-48gn8\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.758486 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp2nk\" (UniqueName: \"kubernetes.io/projected/d960de55-1758-407b-a51d-62fa10fc28dd-kube-api-access-mp2nk\") pod \"nmstate-metrics-5dcf9c57c5-56p9j\" (UID: \"d960de55-1758-407b-a51d-62fa10fc28dd\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.772694 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6"] Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.773495 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.776182 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.777753 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-7csm2" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.778000 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.784139 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6"] Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.859624 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4f03b272-18b2-423e-9e5d-da5fb982df9c-ovs-socket\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.859691 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/72c88d6c-d53a-47e1-a1e2-4b7292c814b4-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-h7t27\" (UID: \"72c88d6c-d53a-47e1-a1e2-4b7292c814b4\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.859748 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4f03b272-18b2-423e-9e5d-da5fb982df9c-dbus-socket\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.859789 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4f03b272-18b2-423e-9e5d-da5fb982df9c-nmstate-lock\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.859800 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4f03b272-18b2-423e-9e5d-da5fb982df9c-ovs-socket\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.859817 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjj2x\" (UniqueName: \"kubernetes.io/projected/72c88d6c-d53a-47e1-a1e2-4b7292c814b4-kube-api-access-fjj2x\") pod \"nmstate-webhook-6b89b748d8-h7t27\" (UID: \"72c88d6c-d53a-47e1-a1e2-4b7292c814b4\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.859982 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48gn8\" (UniqueName: \"kubernetes.io/projected/4f03b272-18b2-423e-9e5d-da5fb982df9c-kube-api-access-48gn8\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.860020 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4f03b272-18b2-423e-9e5d-da5fb982df9c-nmstate-lock\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.860034 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp2nk\" (UniqueName: \"kubernetes.io/projected/d960de55-1758-407b-a51d-62fa10fc28dd-kube-api-access-mp2nk\") pod \"nmstate-metrics-5dcf9c57c5-56p9j\" (UID: \"d960de55-1758-407b-a51d-62fa10fc28dd\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.860269 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4f03b272-18b2-423e-9e5d-da5fb982df9c-dbus-socket\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.887680 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/72c88d6c-d53a-47e1-a1e2-4b7292c814b4-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-h7t27\" (UID: \"72c88d6c-d53a-47e1-a1e2-4b7292c814b4\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.893559 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp2nk\" (UniqueName: \"kubernetes.io/projected/d960de55-1758-407b-a51d-62fa10fc28dd-kube-api-access-mp2nk\") pod \"nmstate-metrics-5dcf9c57c5-56p9j\" (UID: \"d960de55-1758-407b-a51d-62fa10fc28dd\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.897721 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48gn8\" (UniqueName: \"kubernetes.io/projected/4f03b272-18b2-423e-9e5d-da5fb982df9c-kube-api-access-48gn8\") pod \"nmstate-handler-p9xqt\" (UID: \"4f03b272-18b2-423e-9e5d-da5fb982df9c\") " pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.898566 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjj2x\" (UniqueName: \"kubernetes.io/projected/72c88d6c-d53a-47e1-a1e2-4b7292c814b4-kube-api-access-fjj2x\") pod \"nmstate-webhook-6b89b748d8-h7t27\" (UID: \"72c88d6c-d53a-47e1-a1e2-4b7292c814b4\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.909795 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.935276 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.961391 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a5a4df1-9ab8-4b31-b31d-077678cb993f-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-kz4r6\" (UID: \"0a5a4df1-9ab8-4b31-b31d-077678cb993f\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.961470 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0a5a4df1-9ab8-4b31-b31d-077678cb993f-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-kz4r6\" (UID: \"0a5a4df1-9ab8-4b31-b31d-077678cb993f\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.961678 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7w2f\" (UniqueName: \"kubernetes.io/projected/0a5a4df1-9ab8-4b31-b31d-077678cb993f-kube-api-access-t7w2f\") pod \"nmstate-console-plugin-5874bd7bc5-kz4r6\" (UID: \"0a5a4df1-9ab8-4b31-b31d-077678cb993f\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" Nov 24 06:59:13 crc kubenswrapper[4799]: I1124 06:59:13.999271 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7c5b675d7f-6w7w4"] Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.000247 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.015088 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7c5b675d7f-6w7w4"] Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.063946 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a5a4df1-9ab8-4b31-b31d-077678cb993f-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-kz4r6\" (UID: \"0a5a4df1-9ab8-4b31-b31d-077678cb993f\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.063993 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0a5a4df1-9ab8-4b31-b31d-077678cb993f-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-kz4r6\" (UID: \"0a5a4df1-9ab8-4b31-b31d-077678cb993f\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.064046 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7w2f\" (UniqueName: \"kubernetes.io/projected/0a5a4df1-9ab8-4b31-b31d-077678cb993f-kube-api-access-t7w2f\") pod \"nmstate-console-plugin-5874bd7bc5-kz4r6\" (UID: \"0a5a4df1-9ab8-4b31-b31d-077678cb993f\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.071482 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0a5a4df1-9ab8-4b31-b31d-077678cb993f-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-kz4r6\" (UID: \"0a5a4df1-9ab8-4b31-b31d-077678cb993f\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.072441 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0a5a4df1-9ab8-4b31-b31d-077678cb993f-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-kz4r6\" (UID: \"0a5a4df1-9ab8-4b31-b31d-077678cb993f\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.082770 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7w2f\" (UniqueName: \"kubernetes.io/projected/0a5a4df1-9ab8-4b31-b31d-077678cb993f-kube-api-access-t7w2f\") pod \"nmstate-console-plugin-5874bd7bc5-kz4r6\" (UID: \"0a5a4df1-9ab8-4b31-b31d-077678cb993f\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.099267 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.165964 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-trusted-ca-bundle\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.166459 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8xs7\" (UniqueName: \"kubernetes.io/projected/c5608a7b-c770-4354-8b3d-ec21616febcc-kube-api-access-p8xs7\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.166489 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-service-ca\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.166512 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-console-config\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.166538 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5608a7b-c770-4354-8b3d-ec21616febcc-console-serving-cert\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.166635 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5608a7b-c770-4354-8b3d-ec21616febcc-console-oauth-config\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.166678 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-oauth-serving-cert\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.184914 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27"] Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.214344 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j" Nov 24 06:59:14 crc kubenswrapper[4799]: W1124 06:59:14.231337 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72c88d6c_d53a_47e1_a1e2_4b7292c814b4.slice/crio-2585d931337865676931d320584481cfb2373b21aa0cfd93deb065f2617902f7 WatchSource:0}: Error finding container 2585d931337865676931d320584481cfb2373b21aa0cfd93deb065f2617902f7: Status 404 returned error can't find the container with id 2585d931337865676931d320584481cfb2373b21aa0cfd93deb065f2617902f7 Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.267963 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-service-ca\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.268016 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-console-config\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.268033 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5608a7b-c770-4354-8b3d-ec21616febcc-console-serving-cert\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.268102 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5608a7b-c770-4354-8b3d-ec21616febcc-console-oauth-config\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.268133 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-oauth-serving-cert\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.268230 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-trusted-ca-bundle\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.268248 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8xs7\" (UniqueName: \"kubernetes.io/projected/c5608a7b-c770-4354-8b3d-ec21616febcc-kube-api-access-p8xs7\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.269569 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-console-config\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.271203 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-oauth-serving-cert\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.271900 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-service-ca\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.273734 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5608a7b-c770-4354-8b3d-ec21616febcc-trusted-ca-bundle\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.275013 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5608a7b-c770-4354-8b3d-ec21616febcc-console-serving-cert\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.280548 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5608a7b-c770-4354-8b3d-ec21616febcc-console-oauth-config\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.288690 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8xs7\" (UniqueName: \"kubernetes.io/projected/c5608a7b-c770-4354-8b3d-ec21616febcc-kube-api-access-p8xs7\") pod \"console-7c5b675d7f-6w7w4\" (UID: \"c5608a7b-c770-4354-8b3d-ec21616febcc\") " pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.329013 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.346897 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6"] Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.423147 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j"] Nov 24 06:59:14 crc kubenswrapper[4799]: W1124 06:59:14.430994 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd960de55_1758_407b_a51d_62fa10fc28dd.slice/crio-3501a763f04270eb1ae236b200bf1c1e7ee8df71b47cc53dafb591c698ce19bb WatchSource:0}: Error finding container 3501a763f04270eb1ae236b200bf1c1e7ee8df71b47cc53dafb591c698ce19bb: Status 404 returned error can't find the container with id 3501a763f04270eb1ae236b200bf1c1e7ee8df71b47cc53dafb591c698ce19bb Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.554136 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7c5b675d7f-6w7w4"] Nov 24 06:59:14 crc kubenswrapper[4799]: W1124 06:59:14.558562 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5608a7b_c770_4354_8b3d_ec21616febcc.slice/crio-0ad4719987f87c96b7faefb88c97bf8a10c55b5336a4a788c0b57ab731647028 WatchSource:0}: Error finding container 0ad4719987f87c96b7faefb88c97bf8a10c55b5336a4a788c0b57ab731647028: Status 404 returned error can't find the container with id 0ad4719987f87c96b7faefb88c97bf8a10c55b5336a4a788c0b57ab731647028 Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.562521 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-p9xqt" event={"ID":"4f03b272-18b2-423e-9e5d-da5fb982df9c","Type":"ContainerStarted","Data":"d547a28a0a68a950b2f995977c2c8e1c617f123710991f14414d67e78e6e885e"} Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.563878 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j" event={"ID":"d960de55-1758-407b-a51d-62fa10fc28dd","Type":"ContainerStarted","Data":"3501a763f04270eb1ae236b200bf1c1e7ee8df71b47cc53dafb591c698ce19bb"} Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.565005 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" event={"ID":"72c88d6c-d53a-47e1-a1e2-4b7292c814b4","Type":"ContainerStarted","Data":"2585d931337865676931d320584481cfb2373b21aa0cfd93deb065f2617902f7"} Nov 24 06:59:14 crc kubenswrapper[4799]: I1124 06:59:14.566398 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" event={"ID":"0a5a4df1-9ab8-4b31-b31d-077678cb993f","Type":"ContainerStarted","Data":"f939f7aecc32a51dd960f2565fb72061e9940b62b2ec3e638be0bd57e0c4957f"} Nov 24 06:59:15 crc kubenswrapper[4799]: I1124 06:59:15.575872 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7c5b675d7f-6w7w4" event={"ID":"c5608a7b-c770-4354-8b3d-ec21616febcc","Type":"ContainerStarted","Data":"2c5a5c4275e0039a8d0e565a5c227ea488829e7115666e525c6d09279cdc8079"} Nov 24 06:59:15 crc kubenswrapper[4799]: I1124 06:59:15.576417 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7c5b675d7f-6w7w4" event={"ID":"c5608a7b-c770-4354-8b3d-ec21616febcc","Type":"ContainerStarted","Data":"0ad4719987f87c96b7faefb88c97bf8a10c55b5336a4a788c0b57ab731647028"} Nov 24 06:59:15 crc kubenswrapper[4799]: I1124 06:59:15.594089 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7c5b675d7f-6w7w4" podStartSLOduration=2.594067961 podStartE2EDuration="2.594067961s" podCreationTimestamp="2025-11-24 06:59:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 06:59:15.592466236 +0000 UTC m=+701.248448720" watchObservedRunningTime="2025-11-24 06:59:15.594067961 +0000 UTC m=+701.250050435" Nov 24 06:59:17 crc kubenswrapper[4799]: I1124 06:59:17.605184 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" event={"ID":"0a5a4df1-9ab8-4b31-b31d-077678cb993f","Type":"ContainerStarted","Data":"e2cef63e376ba876aeff2825fc3532d62cd666733ace3ccb74f8f1da80155f70"} Nov 24 06:59:17 crc kubenswrapper[4799]: I1124 06:59:17.609196 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-p9xqt" event={"ID":"4f03b272-18b2-423e-9e5d-da5fb982df9c","Type":"ContainerStarted","Data":"191882155ee57f1abe55000487e5337a6def36088187a6f90a31632ea3ef6ec3"} Nov 24 06:59:17 crc kubenswrapper[4799]: I1124 06:59:17.609405 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:17 crc kubenswrapper[4799]: I1124 06:59:17.612208 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j" event={"ID":"d960de55-1758-407b-a51d-62fa10fc28dd","Type":"ContainerStarted","Data":"fe5f7227e3933c48c22b44cd0968bebe4005f3e3121a71669529ad6fb0d79c35"} Nov 24 06:59:17 crc kubenswrapper[4799]: I1124 06:59:17.615815 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" event={"ID":"72c88d6c-d53a-47e1-a1e2-4b7292c814b4","Type":"ContainerStarted","Data":"c9454733cf38843381536cd085b84c50d3aa94b3084bf5af4f06a007e6c989a8"} Nov 24 06:59:17 crc kubenswrapper[4799]: I1124 06:59:17.616006 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" Nov 24 06:59:17 crc kubenswrapper[4799]: I1124 06:59:17.712328 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-kz4r6" podStartSLOduration=2.337636689 podStartE2EDuration="4.712294358s" podCreationTimestamp="2025-11-24 06:59:13 +0000 UTC" firstStartedPulling="2025-11-24 06:59:14.368679031 +0000 UTC m=+700.024661505" lastFinishedPulling="2025-11-24 06:59:16.7433367 +0000 UTC m=+702.399319174" observedRunningTime="2025-11-24 06:59:17.633827948 +0000 UTC m=+703.289810462" watchObservedRunningTime="2025-11-24 06:59:17.712294358 +0000 UTC m=+703.368276852" Nov 24 06:59:17 crc kubenswrapper[4799]: I1124 06:59:17.713038 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-p9xqt" podStartSLOduration=1.944475907 podStartE2EDuration="4.713030748s" podCreationTimestamp="2025-11-24 06:59:13 +0000 UTC" firstStartedPulling="2025-11-24 06:59:13.986305205 +0000 UTC m=+699.642287679" lastFinishedPulling="2025-11-24 06:59:16.754860046 +0000 UTC m=+702.410842520" observedRunningTime="2025-11-24 06:59:17.709799057 +0000 UTC m=+703.365781541" watchObservedRunningTime="2025-11-24 06:59:17.713030748 +0000 UTC m=+703.369013232" Nov 24 06:59:17 crc kubenswrapper[4799]: I1124 06:59:17.734351 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" podStartSLOduration=2.192780222 podStartE2EDuration="4.734324121s" podCreationTimestamp="2025-11-24 06:59:13 +0000 UTC" firstStartedPulling="2025-11-24 06:59:14.235207486 +0000 UTC m=+699.891189960" lastFinishedPulling="2025-11-24 06:59:16.776751385 +0000 UTC m=+702.432733859" observedRunningTime="2025-11-24 06:59:17.733874588 +0000 UTC m=+703.389857082" watchObservedRunningTime="2025-11-24 06:59:17.734324121 +0000 UTC m=+703.390306605" Nov 24 06:59:19 crc kubenswrapper[4799]: I1124 06:59:19.641583 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j" event={"ID":"d960de55-1758-407b-a51d-62fa10fc28dd","Type":"ContainerStarted","Data":"48cff5e2e993f811acdd764a699f99ce7c7cbd90ae9467e15d486a32bcf6c906"} Nov 24 06:59:20 crc kubenswrapper[4799]: I1124 06:59:20.401152 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:59:20 crc kubenswrapper[4799]: I1124 06:59:20.401625 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:59:23 crc kubenswrapper[4799]: I1124 06:59:23.972817 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-p9xqt" Nov 24 06:59:23 crc kubenswrapper[4799]: I1124 06:59:23.993614 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-56p9j" podStartSLOduration=6.710210013 podStartE2EDuration="10.99357897s" podCreationTimestamp="2025-11-24 06:59:13 +0000 UTC" firstStartedPulling="2025-11-24 06:59:14.432723023 +0000 UTC m=+700.088705487" lastFinishedPulling="2025-11-24 06:59:18.71609193 +0000 UTC m=+704.372074444" observedRunningTime="2025-11-24 06:59:19.666840553 +0000 UTC m=+705.322823057" watchObservedRunningTime="2025-11-24 06:59:23.99357897 +0000 UTC m=+709.649561464" Nov 24 06:59:24 crc kubenswrapper[4799]: I1124 06:59:24.330096 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:24 crc kubenswrapper[4799]: I1124 06:59:24.330165 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:24 crc kubenswrapper[4799]: I1124 06:59:24.338396 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:24 crc kubenswrapper[4799]: I1124 06:59:24.705398 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7c5b675d7f-6w7w4" Nov 24 06:59:24 crc kubenswrapper[4799]: I1124 06:59:24.784628 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-58mkm"] Nov 24 06:59:33 crc kubenswrapper[4799]: I1124 06:59:33.919346 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-h7t27" Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.724270 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x"] Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.726277 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.729456 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.745255 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x"] Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.834052 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.834147 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.834180 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpcfq\" (UniqueName: \"kubernetes.io/projected/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-kube-api-access-gpcfq\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.936368 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpcfq\" (UniqueName: \"kubernetes.io/projected/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-kube-api-access-gpcfq\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.936579 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.936665 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.937578 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.937649 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:47 crc kubenswrapper[4799]: I1124 06:59:47.974057 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpcfq\" (UniqueName: \"kubernetes.io/projected/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-kube-api-access-gpcfq\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:48 crc kubenswrapper[4799]: I1124 06:59:48.045687 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:48 crc kubenswrapper[4799]: I1124 06:59:48.268887 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x"] Nov 24 06:59:48 crc kubenswrapper[4799]: I1124 06:59:48.907835 4799 generic.go:334] "Generic (PLEG): container finished" podID="7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" containerID="8631a703ec876e44c0652dfa37c6d393e03a5a3f693d1d32854015ade656fa6a" exitCode=0 Nov 24 06:59:48 crc kubenswrapper[4799]: I1124 06:59:48.908041 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" event={"ID":"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0","Type":"ContainerDied","Data":"8631a703ec876e44c0652dfa37c6d393e03a5a3f693d1d32854015ade656fa6a"} Nov 24 06:59:48 crc kubenswrapper[4799]: I1124 06:59:48.908539 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" event={"ID":"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0","Type":"ContainerStarted","Data":"ab6438e3f7679b61e9744a6765dbd08f930be616ba95d1357623054463094a14"} Nov 24 06:59:49 crc kubenswrapper[4799]: I1124 06:59:49.848658 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-58mkm" podUID="336a419e-e807-4859-b624-2ed06a9a8665" containerName="console" containerID="cri-o://72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7" gracePeriod=15 Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.262170 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-58mkm_336a419e-e807-4859-b624-2ed06a9a8665/console/0.log" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.262304 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.277050 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-console-config\") pod \"336a419e-e807-4859-b624-2ed06a9a8665\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.277758 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-service-ca\") pod \"336a419e-e807-4859-b624-2ed06a9a8665\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.277939 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-oauth-serving-cert\") pod \"336a419e-e807-4859-b624-2ed06a9a8665\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.277970 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-oauth-config\") pod \"336a419e-e807-4859-b624-2ed06a9a8665\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.278081 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-trusted-ca-bundle\") pod \"336a419e-e807-4859-b624-2ed06a9a8665\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.278106 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pqdc\" (UniqueName: \"kubernetes.io/projected/336a419e-e807-4859-b624-2ed06a9a8665-kube-api-access-4pqdc\") pod \"336a419e-e807-4859-b624-2ed06a9a8665\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.278140 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-serving-cert\") pod \"336a419e-e807-4859-b624-2ed06a9a8665\" (UID: \"336a419e-e807-4859-b624-2ed06a9a8665\") " Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.278176 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-console-config" (OuterVolumeSpecName: "console-config") pod "336a419e-e807-4859-b624-2ed06a9a8665" (UID: "336a419e-e807-4859-b624-2ed06a9a8665"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.278767 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-service-ca" (OuterVolumeSpecName: "service-ca") pod "336a419e-e807-4859-b624-2ed06a9a8665" (UID: "336a419e-e807-4859-b624-2ed06a9a8665"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.279335 4799 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.279365 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.279690 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "336a419e-e807-4859-b624-2ed06a9a8665" (UID: "336a419e-e807-4859-b624-2ed06a9a8665"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.279906 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "336a419e-e807-4859-b624-2ed06a9a8665" (UID: "336a419e-e807-4859-b624-2ed06a9a8665"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.289315 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "336a419e-e807-4859-b624-2ed06a9a8665" (UID: "336a419e-e807-4859-b624-2ed06a9a8665"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.302809 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "336a419e-e807-4859-b624-2ed06a9a8665" (UID: "336a419e-e807-4859-b624-2ed06a9a8665"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.303678 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/336a419e-e807-4859-b624-2ed06a9a8665-kube-api-access-4pqdc" (OuterVolumeSpecName: "kube-api-access-4pqdc") pod "336a419e-e807-4859-b624-2ed06a9a8665" (UID: "336a419e-e807-4859-b624-2ed06a9a8665"). InnerVolumeSpecName "kube-api-access-4pqdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.381460 4799 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.381535 4799 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.381560 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/336a419e-e807-4859-b624-2ed06a9a8665-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.381574 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pqdc\" (UniqueName: \"kubernetes.io/projected/336a419e-e807-4859-b624-2ed06a9a8665-kube-api-access-4pqdc\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.381591 4799 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/336a419e-e807-4859-b624-2ed06a9a8665-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.401389 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.401500 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.925296 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-58mkm_336a419e-e807-4859-b624-2ed06a9a8665/console/0.log" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.925365 4799 generic.go:334] "Generic (PLEG): container finished" podID="336a419e-e807-4859-b624-2ed06a9a8665" containerID="72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7" exitCode=2 Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.925445 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-58mkm" event={"ID":"336a419e-e807-4859-b624-2ed06a9a8665","Type":"ContainerDied","Data":"72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7"} Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.925488 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-58mkm" event={"ID":"336a419e-e807-4859-b624-2ed06a9a8665","Type":"ContainerDied","Data":"909b61f4a543f5f2fe4561491241034a8d45e42e797990ba6522d93f744f88ea"} Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.925518 4799 scope.go:117] "RemoveContainer" containerID="72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.925699 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-58mkm" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.933082 4799 generic.go:334] "Generic (PLEG): container finished" podID="7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" containerID="c4b9fc23d0f92e977e48dd420d858c45be3c95b2ea98c6a4ca67aecf0a021486" exitCode=0 Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.933129 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" event={"ID":"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0","Type":"ContainerDied","Data":"c4b9fc23d0f92e977e48dd420d858c45be3c95b2ea98c6a4ca67aecf0a021486"} Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.975004 4799 scope.go:117] "RemoveContainer" containerID="72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7" Nov 24 06:59:50 crc kubenswrapper[4799]: E1124 06:59:50.975827 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7\": container with ID starting with 72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7 not found: ID does not exist" containerID="72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.975959 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7"} err="failed to get container status \"72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7\": rpc error: code = NotFound desc = could not find container \"72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7\": container with ID starting with 72dce300f861261c34763a62f7ab74d31977bef4f0ff3d2919989271908480b7 not found: ID does not exist" Nov 24 06:59:50 crc kubenswrapper[4799]: I1124 06:59:50.986038 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-58mkm"] Nov 24 06:59:51 crc kubenswrapper[4799]: I1124 06:59:51.005189 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-58mkm"] Nov 24 06:59:51 crc kubenswrapper[4799]: I1124 06:59:51.641647 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="336a419e-e807-4859-b624-2ed06a9a8665" path="/var/lib/kubelet/pods/336a419e-e807-4859-b624-2ed06a9a8665/volumes" Nov 24 06:59:51 crc kubenswrapper[4799]: I1124 06:59:51.945078 4799 generic.go:334] "Generic (PLEG): container finished" podID="7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" containerID="0f879f1feb2cc5aa0e4ec679998909991fef097026ebe78cca4b51431a7ee190" exitCode=0 Nov 24 06:59:51 crc kubenswrapper[4799]: I1124 06:59:51.945153 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" event={"ID":"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0","Type":"ContainerDied","Data":"0f879f1feb2cc5aa0e4ec679998909991fef097026ebe78cca4b51431a7ee190"} Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.298445 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.343218 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-util\") pod \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.343328 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpcfq\" (UniqueName: \"kubernetes.io/projected/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-kube-api-access-gpcfq\") pod \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.343372 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-bundle\") pod \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\" (UID: \"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0\") " Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.344466 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-bundle" (OuterVolumeSpecName: "bundle") pod "7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" (UID: "7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.353267 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-kube-api-access-gpcfq" (OuterVolumeSpecName: "kube-api-access-gpcfq") pod "7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" (UID: "7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0"). InnerVolumeSpecName "kube-api-access-gpcfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.365941 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-util" (OuterVolumeSpecName: "util") pod "7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" (UID: "7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.444607 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpcfq\" (UniqueName: \"kubernetes.io/projected/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-kube-api-access-gpcfq\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.444655 4799 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.444674 4799 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0-util\") on node \"crc\" DevicePath \"\"" Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.965512 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" event={"ID":"7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0","Type":"ContainerDied","Data":"ab6438e3f7679b61e9744a6765dbd08f930be616ba95d1357623054463094a14"} Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.965599 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab6438e3f7679b61e9744a6765dbd08f930be616ba95d1357623054463094a14" Nov 24 06:59:53 crc kubenswrapper[4799]: I1124 06:59:53.965725 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.135309 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5"] Nov 24 07:00:00 crc kubenswrapper[4799]: E1124 07:00:00.136231 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" containerName="pull" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.136254 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" containerName="pull" Nov 24 07:00:00 crc kubenswrapper[4799]: E1124 07:00:00.136283 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" containerName="util" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.136294 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" containerName="util" Nov 24 07:00:00 crc kubenswrapper[4799]: E1124 07:00:00.136309 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" containerName="extract" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.136322 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" containerName="extract" Nov 24 07:00:00 crc kubenswrapper[4799]: E1124 07:00:00.136342 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336a419e-e807-4859-b624-2ed06a9a8665" containerName="console" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.136352 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="336a419e-e807-4859-b624-2ed06a9a8665" containerName="console" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.136517 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="336a419e-e807-4859-b624-2ed06a9a8665" containerName="console" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.136537 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0" containerName="extract" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.137213 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.143478 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.143512 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.155128 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5"] Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.244757 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0570ad32-6b7f-44f3-90aa-1cac51750548-secret-volume\") pod \"collect-profiles-29399460-tqpm5\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.244823 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gljvs\" (UniqueName: \"kubernetes.io/projected/0570ad32-6b7f-44f3-90aa-1cac51750548-kube-api-access-gljvs\") pod \"collect-profiles-29399460-tqpm5\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.244865 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0570ad32-6b7f-44f3-90aa-1cac51750548-config-volume\") pod \"collect-profiles-29399460-tqpm5\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.302706 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rzh22"] Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.303079 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" podUID="36f481dd-d3ab-422a-8f26-69fcccb51410" containerName="controller-manager" containerID="cri-o://179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c" gracePeriod=30 Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.345879 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0570ad32-6b7f-44f3-90aa-1cac51750548-secret-volume\") pod \"collect-profiles-29399460-tqpm5\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.346365 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gljvs\" (UniqueName: \"kubernetes.io/projected/0570ad32-6b7f-44f3-90aa-1cac51750548-kube-api-access-gljvs\") pod \"collect-profiles-29399460-tqpm5\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.346398 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0570ad32-6b7f-44f3-90aa-1cac51750548-config-volume\") pod \"collect-profiles-29399460-tqpm5\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.347495 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0570ad32-6b7f-44f3-90aa-1cac51750548-config-volume\") pod \"collect-profiles-29399460-tqpm5\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.356720 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0570ad32-6b7f-44f3-90aa-1cac51750548-secret-volume\") pod \"collect-profiles-29399460-tqpm5\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.371839 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gljvs\" (UniqueName: \"kubernetes.io/projected/0570ad32-6b7f-44f3-90aa-1cac51750548-kube-api-access-gljvs\") pod \"collect-profiles-29399460-tqpm5\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.463783 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.480584 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx"] Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.480934 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" podUID="facf7448-3c40-476b-bc7f-9526c5f4c3e2" containerName="route-controller-manager" containerID="cri-o://89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797" gracePeriod=30 Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.748747 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.855781 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-config\") pod \"36f481dd-d3ab-422a-8f26-69fcccb51410\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.855896 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36f481dd-d3ab-422a-8f26-69fcccb51410-serving-cert\") pod \"36f481dd-d3ab-422a-8f26-69fcccb51410\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.855944 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-proxy-ca-bundles\") pod \"36f481dd-d3ab-422a-8f26-69fcccb51410\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.855976 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gttgg\" (UniqueName: \"kubernetes.io/projected/36f481dd-d3ab-422a-8f26-69fcccb51410-kube-api-access-gttgg\") pod \"36f481dd-d3ab-422a-8f26-69fcccb51410\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.856024 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-client-ca\") pod \"36f481dd-d3ab-422a-8f26-69fcccb51410\" (UID: \"36f481dd-d3ab-422a-8f26-69fcccb51410\") " Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.857026 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-client-ca" (OuterVolumeSpecName: "client-ca") pod "36f481dd-d3ab-422a-8f26-69fcccb51410" (UID: "36f481dd-d3ab-422a-8f26-69fcccb51410"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.857298 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "36f481dd-d3ab-422a-8f26-69fcccb51410" (UID: "36f481dd-d3ab-422a-8f26-69fcccb51410"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.857337 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-config" (OuterVolumeSpecName: "config") pod "36f481dd-d3ab-422a-8f26-69fcccb51410" (UID: "36f481dd-d3ab-422a-8f26-69fcccb51410"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.862560 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36f481dd-d3ab-422a-8f26-69fcccb51410-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "36f481dd-d3ab-422a-8f26-69fcccb51410" (UID: "36f481dd-d3ab-422a-8f26-69fcccb51410"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.862577 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36f481dd-d3ab-422a-8f26-69fcccb51410-kube-api-access-gttgg" (OuterVolumeSpecName: "kube-api-access-gttgg") pod "36f481dd-d3ab-422a-8f26-69fcccb51410" (UID: "36f481dd-d3ab-422a-8f26-69fcccb51410"). InnerVolumeSpecName "kube-api-access-gttgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.874793 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.957080 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.957125 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gttgg\" (UniqueName: \"kubernetes.io/projected/36f481dd-d3ab-422a-8f26-69fcccb51410-kube-api-access-gttgg\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.957136 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.957146 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36f481dd-d3ab-422a-8f26-69fcccb51410-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.957156 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36f481dd-d3ab-422a-8f26-69fcccb51410-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:00 crc kubenswrapper[4799]: I1124 07:00:00.965761 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5"] Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.015180 4799 generic.go:334] "Generic (PLEG): container finished" podID="facf7448-3c40-476b-bc7f-9526c5f4c3e2" containerID="89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797" exitCode=0 Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.015265 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.015277 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" event={"ID":"facf7448-3c40-476b-bc7f-9526c5f4c3e2","Type":"ContainerDied","Data":"89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797"} Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.015330 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx" event={"ID":"facf7448-3c40-476b-bc7f-9526c5f4c3e2","Type":"ContainerDied","Data":"a642fa0b64b17ac7827420f59ca84cdcd610f60612c97e98da8d1347f7b5dc62"} Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.015355 4799 scope.go:117] "RemoveContainer" containerID="89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.017292 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" event={"ID":"0570ad32-6b7f-44f3-90aa-1cac51750548","Type":"ContainerStarted","Data":"7f0efdd63a4985b3bacbe8198d8f8ce89f7897cf7d5d0d69ae6b0c522f65c137"} Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.018657 4799 generic.go:334] "Generic (PLEG): container finished" podID="36f481dd-d3ab-422a-8f26-69fcccb51410" containerID="179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c" exitCode=0 Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.018687 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" event={"ID":"36f481dd-d3ab-422a-8f26-69fcccb51410","Type":"ContainerDied","Data":"179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c"} Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.018734 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" event={"ID":"36f481dd-d3ab-422a-8f26-69fcccb51410","Type":"ContainerDied","Data":"0034ea623d86f7e724d60cf08ff7e83b6031880f026455483c99c21fd0951459"} Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.018761 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rzh22" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.044791 4799 scope.go:117] "RemoveContainer" containerID="89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797" Nov 24 07:00:01 crc kubenswrapper[4799]: E1124 07:00:01.045472 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797\": container with ID starting with 89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797 not found: ID does not exist" containerID="89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.045538 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797"} err="failed to get container status \"89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797\": rpc error: code = NotFound desc = could not find container \"89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797\": container with ID starting with 89947740c49d7c22ec2d319c43d0f5c970e6ccd1f9b9a9df3590b364d7836797 not found: ID does not exist" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.045574 4799 scope.go:117] "RemoveContainer" containerID="179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.057707 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-client-ca\") pod \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.057823 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg4jq\" (UniqueName: \"kubernetes.io/projected/facf7448-3c40-476b-bc7f-9526c5f4c3e2-kube-api-access-kg4jq\") pod \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.057962 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-config\") pod \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.058003 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/facf7448-3c40-476b-bc7f-9526c5f4c3e2-serving-cert\") pod \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\" (UID: \"facf7448-3c40-476b-bc7f-9526c5f4c3e2\") " Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.058653 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-client-ca" (OuterVolumeSpecName: "client-ca") pod "facf7448-3c40-476b-bc7f-9526c5f4c3e2" (UID: "facf7448-3c40-476b-bc7f-9526c5f4c3e2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.058682 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-config" (OuterVolumeSpecName: "config") pod "facf7448-3c40-476b-bc7f-9526c5f4c3e2" (UID: "facf7448-3c40-476b-bc7f-9526c5f4c3e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.061928 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/facf7448-3c40-476b-bc7f-9526c5f4c3e2-kube-api-access-kg4jq" (OuterVolumeSpecName: "kube-api-access-kg4jq") pod "facf7448-3c40-476b-bc7f-9526c5f4c3e2" (UID: "facf7448-3c40-476b-bc7f-9526c5f4c3e2"). InnerVolumeSpecName "kube-api-access-kg4jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.065760 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/facf7448-3c40-476b-bc7f-9526c5f4c3e2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "facf7448-3c40-476b-bc7f-9526c5f4c3e2" (UID: "facf7448-3c40-476b-bc7f-9526c5f4c3e2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.070977 4799 scope.go:117] "RemoveContainer" containerID="179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c" Nov 24 07:00:01 crc kubenswrapper[4799]: E1124 07:00:01.071505 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c\": container with ID starting with 179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c not found: ID does not exist" containerID="179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.071579 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c"} err="failed to get container status \"179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c\": rpc error: code = NotFound desc = could not find container \"179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c\": container with ID starting with 179a5b1b27887e56d6cc1494e3c29e90ac26ce3661bc0dc2d27d2acf93babf0c not found: ID does not exist" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.072828 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rzh22"] Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.076615 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rzh22"] Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.159378 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg4jq\" (UniqueName: \"kubernetes.io/projected/facf7448-3c40-476b-bc7f-9526c5f4c3e2-kube-api-access-kg4jq\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.159901 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.159916 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/facf7448-3c40-476b-bc7f-9526c5f4c3e2-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.159932 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/facf7448-3c40-476b-bc7f-9526c5f4c3e2-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.357883 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx"] Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.360784 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-77hvx"] Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.636338 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36f481dd-d3ab-422a-8f26-69fcccb51410" path="/var/lib/kubelet/pods/36f481dd-d3ab-422a-8f26-69fcccb51410/volumes" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.637674 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="facf7448-3c40-476b-bc7f-9526c5f4c3e2" path="/var/lib/kubelet/pods/facf7448-3c40-476b-bc7f-9526c5f4c3e2/volumes" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.743994 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-dcd74bdbc-b26jr"] Nov 24 07:00:01 crc kubenswrapper[4799]: E1124 07:00:01.744287 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36f481dd-d3ab-422a-8f26-69fcccb51410" containerName="controller-manager" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.744305 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="36f481dd-d3ab-422a-8f26-69fcccb51410" containerName="controller-manager" Nov 24 07:00:01 crc kubenswrapper[4799]: E1124 07:00:01.744323 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="facf7448-3c40-476b-bc7f-9526c5f4c3e2" containerName="route-controller-manager" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.744329 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="facf7448-3c40-476b-bc7f-9526c5f4c3e2" containerName="route-controller-manager" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.744432 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="36f481dd-d3ab-422a-8f26-69fcccb51410" containerName="controller-manager" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.744447 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="facf7448-3c40-476b-bc7f-9526c5f4c3e2" containerName="route-controller-manager" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.744904 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.748204 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.750417 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.750570 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.750588 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.750718 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.750943 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.764156 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.767306 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p"] Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.768680 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.770885 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.771761 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.772714 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.772926 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.778890 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.778890 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.780837 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-dcd74bdbc-b26jr"] Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.790147 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p"] Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.869352 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c7cc88ec-a737-4473-b310-a0deb997f04a-proxy-ca-bundles\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.869456 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kkwl\" (UniqueName: \"kubernetes.io/projected/6441687e-f4c2-4500-8b17-b90f8414b2ba-kube-api-access-7kkwl\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.869517 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6441687e-f4c2-4500-8b17-b90f8414b2ba-client-ca\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.869562 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7cc88ec-a737-4473-b310-a0deb997f04a-config\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.869588 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6441687e-f4c2-4500-8b17-b90f8414b2ba-config\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.869608 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7cc88ec-a737-4473-b310-a0deb997f04a-client-ca\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.869694 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2wgt\" (UniqueName: \"kubernetes.io/projected/c7cc88ec-a737-4473-b310-a0deb997f04a-kube-api-access-t2wgt\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.869789 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6441687e-f4c2-4500-8b17-b90f8414b2ba-serving-cert\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.869811 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7cc88ec-a737-4473-b310-a0deb997f04a-serving-cert\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.970868 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kkwl\" (UniqueName: \"kubernetes.io/projected/6441687e-f4c2-4500-8b17-b90f8414b2ba-kube-api-access-7kkwl\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.970960 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6441687e-f4c2-4500-8b17-b90f8414b2ba-client-ca\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.970991 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7cc88ec-a737-4473-b310-a0deb997f04a-config\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.971017 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6441687e-f4c2-4500-8b17-b90f8414b2ba-config\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.971039 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7cc88ec-a737-4473-b310-a0deb997f04a-client-ca\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.971073 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2wgt\" (UniqueName: \"kubernetes.io/projected/c7cc88ec-a737-4473-b310-a0deb997f04a-kube-api-access-t2wgt\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.971110 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6441687e-f4c2-4500-8b17-b90f8414b2ba-serving-cert\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.971133 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7cc88ec-a737-4473-b310-a0deb997f04a-serving-cert\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.971371 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c7cc88ec-a737-4473-b310-a0deb997f04a-proxy-ca-bundles\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.972519 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7cc88ec-a737-4473-b310-a0deb997f04a-client-ca\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.972775 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c7cc88ec-a737-4473-b310-a0deb997f04a-proxy-ca-bundles\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.973769 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6441687e-f4c2-4500-8b17-b90f8414b2ba-config\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.973805 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6441687e-f4c2-4500-8b17-b90f8414b2ba-client-ca\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.974633 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7cc88ec-a737-4473-b310-a0deb997f04a-config\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.978310 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7cc88ec-a737-4473-b310-a0deb997f04a-serving-cert\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.986451 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6441687e-f4c2-4500-8b17-b90f8414b2ba-serving-cert\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.988250 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2wgt\" (UniqueName: \"kubernetes.io/projected/c7cc88ec-a737-4473-b310-a0deb997f04a-kube-api-access-t2wgt\") pod \"controller-manager-dcd74bdbc-b26jr\" (UID: \"c7cc88ec-a737-4473-b310-a0deb997f04a\") " pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:01 crc kubenswrapper[4799]: I1124 07:00:01.993624 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kkwl\" (UniqueName: \"kubernetes.io/projected/6441687e-f4c2-4500-8b17-b90f8414b2ba-kube-api-access-7kkwl\") pod \"route-controller-manager-65f667c589-fcg9p\" (UID: \"6441687e-f4c2-4500-8b17-b90f8414b2ba\") " pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.027138 4799 generic.go:334] "Generic (PLEG): container finished" podID="0570ad32-6b7f-44f3-90aa-1cac51750548" containerID="8682d1cf7ee42852f953abfeecee75629e7ea60dc5dc60582f4a0c0b778ff7fa" exitCode=0 Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.027196 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" event={"ID":"0570ad32-6b7f-44f3-90aa-1cac51750548","Type":"ContainerDied","Data":"8682d1cf7ee42852f953abfeecee75629e7ea60dc5dc60582f4a0c0b778ff7fa"} Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.064219 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.081887 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.397413 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p"] Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.530719 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-dcd74bdbc-b26jr"] Nov 24 07:00:02 crc kubenswrapper[4799]: W1124 07:00:02.536762 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7cc88ec_a737_4473_b310_a0deb997f04a.slice/crio-20eee70ee89438f7aacc0f9b92ad12df9dda00da2636f7cac601a13791eda281 WatchSource:0}: Error finding container 20eee70ee89438f7aacc0f9b92ad12df9dda00da2636f7cac601a13791eda281: Status 404 returned error can't find the container with id 20eee70ee89438f7aacc0f9b92ad12df9dda00da2636f7cac601a13791eda281 Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.849184 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx"] Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.850981 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.864863 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.865191 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.865298 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.865237 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.873121 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx"] Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.985459 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jldnk\" (UniqueName: \"kubernetes.io/projected/ead0c22c-848d-4391-903b-62aa4bbc697c-kube-api-access-jldnk\") pod \"metallb-operator-controller-manager-66f4d4788d-6qmmx\" (UID: \"ead0c22c-848d-4391-903b-62aa4bbc697c\") " pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.985568 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ead0c22c-848d-4391-903b-62aa4bbc697c-webhook-cert\") pod \"metallb-operator-controller-manager-66f4d4788d-6qmmx\" (UID: \"ead0c22c-848d-4391-903b-62aa4bbc697c\") " pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:02 crc kubenswrapper[4799]: I1124 07:00:02.985615 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ead0c22c-848d-4391-903b-62aa4bbc697c-apiservice-cert\") pod \"metallb-operator-controller-manager-66f4d4788d-6qmmx\" (UID: \"ead0c22c-848d-4391-903b-62aa4bbc697c\") " pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.035407 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" event={"ID":"c7cc88ec-a737-4473-b310-a0deb997f04a","Type":"ContainerStarted","Data":"d2d256334e556bfdf405983ef96e9039e753478ed5e8b2d373568e35c28e2dcb"} Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.035479 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" event={"ID":"c7cc88ec-a737-4473-b310-a0deb997f04a","Type":"ContainerStarted","Data":"20eee70ee89438f7aacc0f9b92ad12df9dda00da2636f7cac601a13791eda281"} Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.035717 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.037285 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" event={"ID":"6441687e-f4c2-4500-8b17-b90f8414b2ba","Type":"ContainerStarted","Data":"d91813c9b160b94569896c1d4225dffac8a63fbe46e9750d938b807c86d9ed8a"} Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.037340 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" event={"ID":"6441687e-f4c2-4500-8b17-b90f8414b2ba","Type":"ContainerStarted","Data":"cade6464bc3bf391c10f0e0c78d090582e58ddf86c2e3c177668bb995b266279"} Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.068461 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" podStartSLOduration=2.068429428 podStartE2EDuration="2.068429428s" podCreationTimestamp="2025-11-24 07:00:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:00:03.067748179 +0000 UTC m=+748.723730653" watchObservedRunningTime="2025-11-24 07:00:03.068429428 +0000 UTC m=+748.724411902" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.071406 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-dcd74bdbc-b26jr" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.086701 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ead0c22c-848d-4391-903b-62aa4bbc697c-webhook-cert\") pod \"metallb-operator-controller-manager-66f4d4788d-6qmmx\" (UID: \"ead0c22c-848d-4391-903b-62aa4bbc697c\") " pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.087019 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ead0c22c-848d-4391-903b-62aa4bbc697c-apiservice-cert\") pod \"metallb-operator-controller-manager-66f4d4788d-6qmmx\" (UID: \"ead0c22c-848d-4391-903b-62aa4bbc697c\") " pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.087151 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jldnk\" (UniqueName: \"kubernetes.io/projected/ead0c22c-848d-4391-903b-62aa4bbc697c-kube-api-access-jldnk\") pod \"metallb-operator-controller-manager-66f4d4788d-6qmmx\" (UID: \"ead0c22c-848d-4391-903b-62aa4bbc697c\") " pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.097210 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ead0c22c-848d-4391-903b-62aa4bbc697c-webhook-cert\") pod \"metallb-operator-controller-manager-66f4d4788d-6qmmx\" (UID: \"ead0c22c-848d-4391-903b-62aa4bbc697c\") " pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.116035 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" podStartSLOduration=2.115995023 podStartE2EDuration="2.115995023s" podCreationTimestamp="2025-11-24 07:00:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:00:03.115917721 +0000 UTC m=+748.771900195" watchObservedRunningTime="2025-11-24 07:00:03.115995023 +0000 UTC m=+748.771977497" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.125153 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ead0c22c-848d-4391-903b-62aa4bbc697c-apiservice-cert\") pod \"metallb-operator-controller-manager-66f4d4788d-6qmmx\" (UID: \"ead0c22c-848d-4391-903b-62aa4bbc697c\") " pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.131484 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jldnk\" (UniqueName: \"kubernetes.io/projected/ead0c22c-848d-4391-903b-62aa4bbc697c-kube-api-access-jldnk\") pod \"metallb-operator-controller-manager-66f4d4788d-6qmmx\" (UID: \"ead0c22c-848d-4391-903b-62aa4bbc697c\") " pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.190515 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.388655 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.406766 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp"] Nov 24 07:00:03 crc kubenswrapper[4799]: E1124 07:00:03.410299 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0570ad32-6b7f-44f3-90aa-1cac51750548" containerName="collect-profiles" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.410338 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0570ad32-6b7f-44f3-90aa-1cac51750548" containerName="collect-profiles" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.410452 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0570ad32-6b7f-44f3-90aa-1cac51750548" containerName="collect-profiles" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.412183 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.418458 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.425882 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.426622 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-74kmg" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.460994 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp"] Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.495884 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gljvs\" (UniqueName: \"kubernetes.io/projected/0570ad32-6b7f-44f3-90aa-1cac51750548-kube-api-access-gljvs\") pod \"0570ad32-6b7f-44f3-90aa-1cac51750548\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.496089 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0570ad32-6b7f-44f3-90aa-1cac51750548-secret-volume\") pod \"0570ad32-6b7f-44f3-90aa-1cac51750548\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.496143 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0570ad32-6b7f-44f3-90aa-1cac51750548-config-volume\") pod \"0570ad32-6b7f-44f3-90aa-1cac51750548\" (UID: \"0570ad32-6b7f-44f3-90aa-1cac51750548\") " Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.497297 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0570ad32-6b7f-44f3-90aa-1cac51750548-config-volume" (OuterVolumeSpecName: "config-volume") pod "0570ad32-6b7f-44f3-90aa-1cac51750548" (UID: "0570ad32-6b7f-44f3-90aa-1cac51750548"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.502038 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0570ad32-6b7f-44f3-90aa-1cac51750548-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0570ad32-6b7f-44f3-90aa-1cac51750548" (UID: "0570ad32-6b7f-44f3-90aa-1cac51750548"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.517350 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0570ad32-6b7f-44f3-90aa-1cac51750548-kube-api-access-gljvs" (OuterVolumeSpecName: "kube-api-access-gljvs") pod "0570ad32-6b7f-44f3-90aa-1cac51750548" (UID: "0570ad32-6b7f-44f3-90aa-1cac51750548"). InnerVolumeSpecName "kube-api-access-gljvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.601657 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5fcab455-bba7-4f3c-84c7-8e938d1ce6cd-webhook-cert\") pod \"metallb-operator-webhook-server-666db98b56-rbfxp\" (UID: \"5fcab455-bba7-4f3c-84c7-8e938d1ce6cd\") " pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.601725 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5fcab455-bba7-4f3c-84c7-8e938d1ce6cd-apiservice-cert\") pod \"metallb-operator-webhook-server-666db98b56-rbfxp\" (UID: \"5fcab455-bba7-4f3c-84c7-8e938d1ce6cd\") " pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.601759 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8572\" (UniqueName: \"kubernetes.io/projected/5fcab455-bba7-4f3c-84c7-8e938d1ce6cd-kube-api-access-n8572\") pod \"metallb-operator-webhook-server-666db98b56-rbfxp\" (UID: \"5fcab455-bba7-4f3c-84c7-8e938d1ce6cd\") " pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.601807 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gljvs\" (UniqueName: \"kubernetes.io/projected/0570ad32-6b7f-44f3-90aa-1cac51750548-kube-api-access-gljvs\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.601824 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0570ad32-6b7f-44f3-90aa-1cac51750548-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.601834 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0570ad32-6b7f-44f3-90aa-1cac51750548-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.706610 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5fcab455-bba7-4f3c-84c7-8e938d1ce6cd-apiservice-cert\") pod \"metallb-operator-webhook-server-666db98b56-rbfxp\" (UID: \"5fcab455-bba7-4f3c-84c7-8e938d1ce6cd\") " pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.706681 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8572\" (UniqueName: \"kubernetes.io/projected/5fcab455-bba7-4f3c-84c7-8e938d1ce6cd-kube-api-access-n8572\") pod \"metallb-operator-webhook-server-666db98b56-rbfxp\" (UID: \"5fcab455-bba7-4f3c-84c7-8e938d1ce6cd\") " pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.706756 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5fcab455-bba7-4f3c-84c7-8e938d1ce6cd-webhook-cert\") pod \"metallb-operator-webhook-server-666db98b56-rbfxp\" (UID: \"5fcab455-bba7-4f3c-84c7-8e938d1ce6cd\") " pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.720787 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5fcab455-bba7-4f3c-84c7-8e938d1ce6cd-apiservice-cert\") pod \"metallb-operator-webhook-server-666db98b56-rbfxp\" (UID: \"5fcab455-bba7-4f3c-84c7-8e938d1ce6cd\") " pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.722264 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5fcab455-bba7-4f3c-84c7-8e938d1ce6cd-webhook-cert\") pod \"metallb-operator-webhook-server-666db98b56-rbfxp\" (UID: \"5fcab455-bba7-4f3c-84c7-8e938d1ce6cd\") " pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.779004 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8572\" (UniqueName: \"kubernetes.io/projected/5fcab455-bba7-4f3c-84c7-8e938d1ce6cd-kube-api-access-n8572\") pod \"metallb-operator-webhook-server-666db98b56-rbfxp\" (UID: \"5fcab455-bba7-4f3c-84c7-8e938d1ce6cd\") " pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:03 crc kubenswrapper[4799]: I1124 07:00:03.819201 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx"] Nov 24 07:00:03 crc kubenswrapper[4799]: W1124 07:00:03.837222 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podead0c22c_848d_4391_903b_62aa4bbc697c.slice/crio-f989ca1d595f7d1cb7a6a57b8e52ac283ce7db92642d82059cc7e61801c2cca1 WatchSource:0}: Error finding container f989ca1d595f7d1cb7a6a57b8e52ac283ce7db92642d82059cc7e61801c2cca1: Status 404 returned error can't find the container with id f989ca1d595f7d1cb7a6a57b8e52ac283ce7db92642d82059cc7e61801c2cca1 Nov 24 07:00:04 crc kubenswrapper[4799]: I1124 07:00:04.046935 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" event={"ID":"0570ad32-6b7f-44f3-90aa-1cac51750548","Type":"ContainerDied","Data":"7f0efdd63a4985b3bacbe8198d8f8ce89f7897cf7d5d0d69ae6b0c522f65c137"} Nov 24 07:00:04 crc kubenswrapper[4799]: I1124 07:00:04.046992 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f0efdd63a4985b3bacbe8198d8f8ce89f7897cf7d5d0d69ae6b0c522f65c137" Nov 24 07:00:04 crc kubenswrapper[4799]: I1124 07:00:04.047059 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5" Nov 24 07:00:04 crc kubenswrapper[4799]: I1124 07:00:04.049545 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" event={"ID":"ead0c22c-848d-4391-903b-62aa4bbc697c","Type":"ContainerStarted","Data":"f989ca1d595f7d1cb7a6a57b8e52ac283ce7db92642d82059cc7e61801c2cca1"} Nov 24 07:00:04 crc kubenswrapper[4799]: I1124 07:00:04.050290 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:04 crc kubenswrapper[4799]: I1124 07:00:04.059271 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-65f667c589-fcg9p" Nov 24 07:00:04 crc kubenswrapper[4799]: I1124 07:00:04.079551 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:04 crc kubenswrapper[4799]: I1124 07:00:04.524902 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp"] Nov 24 07:00:04 crc kubenswrapper[4799]: W1124 07:00:04.530786 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fcab455_bba7_4f3c_84c7_8e938d1ce6cd.slice/crio-b2c061561a462dc1bf222ef7d0d3f962ec30a99e285e0cce1b9d163c3d14e804 WatchSource:0}: Error finding container b2c061561a462dc1bf222ef7d0d3f962ec30a99e285e0cce1b9d163c3d14e804: Status 404 returned error can't find the container with id b2c061561a462dc1bf222ef7d0d3f962ec30a99e285e0cce1b9d163c3d14e804 Nov 24 07:00:05 crc kubenswrapper[4799]: I1124 07:00:05.060918 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" event={"ID":"5fcab455-bba7-4f3c-84c7-8e938d1ce6cd","Type":"ContainerStarted","Data":"b2c061561a462dc1bf222ef7d0d3f962ec30a99e285e0cce1b9d163c3d14e804"} Nov 24 07:00:08 crc kubenswrapper[4799]: I1124 07:00:08.097011 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" event={"ID":"ead0c22c-848d-4391-903b-62aa4bbc697c","Type":"ContainerStarted","Data":"b59e4612f64f4a7d1c22bab08ecac5aaa9ec11cc7e31af97b3c378fa4e022cd9"} Nov 24 07:00:08 crc kubenswrapper[4799]: I1124 07:00:08.098042 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:08 crc kubenswrapper[4799]: I1124 07:00:08.125152 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" podStartSLOduration=2.622683392 podStartE2EDuration="6.125129391s" podCreationTimestamp="2025-11-24 07:00:02 +0000 UTC" firstStartedPulling="2025-11-24 07:00:03.842135393 +0000 UTC m=+749.498117867" lastFinishedPulling="2025-11-24 07:00:07.344581392 +0000 UTC m=+753.000563866" observedRunningTime="2025-11-24 07:00:08.118962386 +0000 UTC m=+753.774944880" watchObservedRunningTime="2025-11-24 07:00:08.125129391 +0000 UTC m=+753.781111865" Nov 24 07:00:10 crc kubenswrapper[4799]: I1124 07:00:10.127908 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" event={"ID":"5fcab455-bba7-4f3c-84c7-8e938d1ce6cd","Type":"ContainerStarted","Data":"9317be521ffb514c880056608a1943f97ef33b861a46ba1e824e896f50d2f8d0"} Nov 24 07:00:10 crc kubenswrapper[4799]: I1124 07:00:10.128519 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:10 crc kubenswrapper[4799]: I1124 07:00:10.148346 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" podStartSLOduration=2.558645286 podStartE2EDuration="7.148321138s" podCreationTimestamp="2025-11-24 07:00:03 +0000 UTC" firstStartedPulling="2025-11-24 07:00:04.533465068 +0000 UTC m=+750.189447542" lastFinishedPulling="2025-11-24 07:00:09.12314091 +0000 UTC m=+754.779123394" observedRunningTime="2025-11-24 07:00:10.146314401 +0000 UTC m=+755.802296885" watchObservedRunningTime="2025-11-24 07:00:10.148321138 +0000 UTC m=+755.804303622" Nov 24 07:00:10 crc kubenswrapper[4799]: I1124 07:00:10.390521 4799 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 07:00:20 crc kubenswrapper[4799]: I1124 07:00:20.400774 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:00:20 crc kubenswrapper[4799]: I1124 07:00:20.401722 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:00:20 crc kubenswrapper[4799]: I1124 07:00:20.401793 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:00:20 crc kubenswrapper[4799]: I1124 07:00:20.402559 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d29a0547b29f459c7d7db9037c62a8945e7f6c626531ca09a0a9fd2c39286a5a"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:00:20 crc kubenswrapper[4799]: I1124 07:00:20.402647 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://d29a0547b29f459c7d7db9037c62a8945e7f6c626531ca09a0a9fd2c39286a5a" gracePeriod=600 Nov 24 07:00:21 crc kubenswrapper[4799]: I1124 07:00:21.202229 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="d29a0547b29f459c7d7db9037c62a8945e7f6c626531ca09a0a9fd2c39286a5a" exitCode=0 Nov 24 07:00:21 crc kubenswrapper[4799]: I1124 07:00:21.202473 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"d29a0547b29f459c7d7db9037c62a8945e7f6c626531ca09a0a9fd2c39286a5a"} Nov 24 07:00:21 crc kubenswrapper[4799]: I1124 07:00:21.203185 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"985590a83c847e8770acbc79642e467b641116d42bb23156312a6fcbb0de12d9"} Nov 24 07:00:21 crc kubenswrapper[4799]: I1124 07:00:21.203222 4799 scope.go:117] "RemoveContainer" containerID="1155080c324c74d531908515f79c186dcf0bbbcf9d2021dbf86a19ed644ed48b" Nov 24 07:00:24 crc kubenswrapper[4799]: I1124 07:00:24.088010 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-666db98b56-rbfxp" Nov 24 07:00:43 crc kubenswrapper[4799]: I1124 07:00:43.194031 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-66f4d4788d-6qmmx" Nov 24 07:00:43 crc kubenswrapper[4799]: I1124 07:00:43.958496 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-7khps"] Nov 24 07:00:43 crc kubenswrapper[4799]: I1124 07:00:43.960030 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" Nov 24 07:00:43 crc kubenswrapper[4799]: I1124 07:00:43.963350 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-2gqq5"] Nov 24 07:00:43 crc kubenswrapper[4799]: I1124 07:00:43.964938 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 07:00:43 crc kubenswrapper[4799]: I1124 07:00:43.965054 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-cd424" Nov 24 07:00:43 crc kubenswrapper[4799]: I1124 07:00:43.972828 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:43 crc kubenswrapper[4799]: I1124 07:00:43.977332 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 07:00:43 crc kubenswrapper[4799]: I1124 07:00:43.981453 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 07:00:43 crc kubenswrapper[4799]: I1124 07:00:43.982209 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-7khps"] Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.056181 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-metrics\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.056249 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrvkj\" (UniqueName: \"kubernetes.io/projected/aa96ad1b-f53e-428e-ac55-bb15c9740c26-kube-api-access-wrvkj\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.056314 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-frr-conf\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.056487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6cd58852-5a3f-46aa-b43a-6a8ab49711a5-cert\") pod \"frr-k8s-webhook-server-6998585d5-7khps\" (UID: \"6cd58852-5a3f-46aa-b43a-6a8ab49711a5\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.056595 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa96ad1b-f53e-428e-ac55-bb15c9740c26-metrics-certs\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.056659 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/aa96ad1b-f53e-428e-ac55-bb15c9740c26-frr-startup\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.056711 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-frr-sockets\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.056897 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-reloader\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.057008 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6w8m\" (UniqueName: \"kubernetes.io/projected/6cd58852-5a3f-46aa-b43a-6a8ab49711a5-kube-api-access-r6w8m\") pod \"frr-k8s-webhook-server-6998585d5-7khps\" (UID: \"6cd58852-5a3f-46aa-b43a-6a8ab49711a5\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.076481 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-bzp2s"] Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.077520 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.080427 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.080733 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.081210 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.082008 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-9p2d4" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.101353 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-gkkl2"] Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.102692 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.104591 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.122540 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-gkkl2"] Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158107 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrvkj\" (UniqueName: \"kubernetes.io/projected/aa96ad1b-f53e-428e-ac55-bb15c9740c26-kube-api-access-wrvkj\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158160 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14418811-1eac-4ac6-a598-8bde1dc50a82-cert\") pod \"controller-6c7b4b5f48-gkkl2\" (UID: \"14418811-1eac-4ac6-a598-8bde1dc50a82\") " pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158185 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-memberlist\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158208 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-frr-conf\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158246 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6cd58852-5a3f-46aa-b43a-6a8ab49711a5-cert\") pod \"frr-k8s-webhook-server-6998585d5-7khps\" (UID: \"6cd58852-5a3f-46aa-b43a-6a8ab49711a5\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158266 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-metallb-excludel2\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158291 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa96ad1b-f53e-428e-ac55-bb15c9740c26-metrics-certs\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158316 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/aa96ad1b-f53e-428e-ac55-bb15c9740c26-frr-startup\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158337 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-frr-sockets\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158356 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/14418811-1eac-4ac6-a598-8bde1dc50a82-metrics-certs\") pod \"controller-6c7b4b5f48-gkkl2\" (UID: \"14418811-1eac-4ac6-a598-8bde1dc50a82\") " pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158387 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-reloader\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158410 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tlmh\" (UniqueName: \"kubernetes.io/projected/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-kube-api-access-6tlmh\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158433 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-metrics-certs\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158454 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6w8m\" (UniqueName: \"kubernetes.io/projected/6cd58852-5a3f-46aa-b43a-6a8ab49711a5-kube-api-access-r6w8m\") pod \"frr-k8s-webhook-server-6998585d5-7khps\" (UID: \"6cd58852-5a3f-46aa-b43a-6a8ab49711a5\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158477 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-metrics\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.158492 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cncwv\" (UniqueName: \"kubernetes.io/projected/14418811-1eac-4ac6-a598-8bde1dc50a82-kube-api-access-cncwv\") pod \"controller-6c7b4b5f48-gkkl2\" (UID: \"14418811-1eac-4ac6-a598-8bde1dc50a82\") " pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.159463 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-frr-sockets\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.159660 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-metrics\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.159886 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-reloader\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.160324 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/aa96ad1b-f53e-428e-ac55-bb15c9740c26-frr-startup\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.160521 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/aa96ad1b-f53e-428e-ac55-bb15c9740c26-frr-conf\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.169838 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa96ad1b-f53e-428e-ac55-bb15c9740c26-metrics-certs\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.173591 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6cd58852-5a3f-46aa-b43a-6a8ab49711a5-cert\") pod \"frr-k8s-webhook-server-6998585d5-7khps\" (UID: \"6cd58852-5a3f-46aa-b43a-6a8ab49711a5\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.181964 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrvkj\" (UniqueName: \"kubernetes.io/projected/aa96ad1b-f53e-428e-ac55-bb15c9740c26-kube-api-access-wrvkj\") pod \"frr-k8s-2gqq5\" (UID: \"aa96ad1b-f53e-428e-ac55-bb15c9740c26\") " pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.194662 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6w8m\" (UniqueName: \"kubernetes.io/projected/6cd58852-5a3f-46aa-b43a-6a8ab49711a5-kube-api-access-r6w8m\") pod \"frr-k8s-webhook-server-6998585d5-7khps\" (UID: \"6cd58852-5a3f-46aa-b43a-6a8ab49711a5\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.260023 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/14418811-1eac-4ac6-a598-8bde1dc50a82-metrics-certs\") pod \"controller-6c7b4b5f48-gkkl2\" (UID: \"14418811-1eac-4ac6-a598-8bde1dc50a82\") " pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.260108 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tlmh\" (UniqueName: \"kubernetes.io/projected/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-kube-api-access-6tlmh\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.260148 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-metrics-certs\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.260192 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cncwv\" (UniqueName: \"kubernetes.io/projected/14418811-1eac-4ac6-a598-8bde1dc50a82-kube-api-access-cncwv\") pod \"controller-6c7b4b5f48-gkkl2\" (UID: \"14418811-1eac-4ac6-a598-8bde1dc50a82\") " pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.260220 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14418811-1eac-4ac6-a598-8bde1dc50a82-cert\") pod \"controller-6c7b4b5f48-gkkl2\" (UID: \"14418811-1eac-4ac6-a598-8bde1dc50a82\") " pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.260243 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-memberlist\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.260294 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-metallb-excludel2\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.261219 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-metallb-excludel2\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: E1124 07:00:44.261416 4799 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 07:00:44 crc kubenswrapper[4799]: E1124 07:00:44.261490 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-memberlist podName:8a7ba0ee-de14-43e5-8b93-9b915e8d9938 nodeName:}" failed. No retries permitted until 2025-11-24 07:00:44.761465304 +0000 UTC m=+790.417447798 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-memberlist") pod "speaker-bzp2s" (UID: "8a7ba0ee-de14-43e5-8b93-9b915e8d9938") : secret "metallb-memberlist" not found Nov 24 07:00:44 crc kubenswrapper[4799]: E1124 07:00:44.261731 4799 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 24 07:00:44 crc kubenswrapper[4799]: E1124 07:00:44.261827 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-metrics-certs podName:8a7ba0ee-de14-43e5-8b93-9b915e8d9938 nodeName:}" failed. No retries permitted until 2025-11-24 07:00:44.761818024 +0000 UTC m=+790.417800498 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-metrics-certs") pod "speaker-bzp2s" (UID: "8a7ba0ee-de14-43e5-8b93-9b915e8d9938") : secret "speaker-certs-secret" not found Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.263120 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.264663 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/14418811-1eac-4ac6-a598-8bde1dc50a82-metrics-certs\") pod \"controller-6c7b4b5f48-gkkl2\" (UID: \"14418811-1eac-4ac6-a598-8bde1dc50a82\") " pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.280674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14418811-1eac-4ac6-a598-8bde1dc50a82-cert\") pod \"controller-6c7b4b5f48-gkkl2\" (UID: \"14418811-1eac-4ac6-a598-8bde1dc50a82\") " pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.288308 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.291061 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cncwv\" (UniqueName: \"kubernetes.io/projected/14418811-1eac-4ac6-a598-8bde1dc50a82-kube-api-access-cncwv\") pod \"controller-6c7b4b5f48-gkkl2\" (UID: \"14418811-1eac-4ac6-a598-8bde1dc50a82\") " pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.292517 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.300762 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tlmh\" (UniqueName: \"kubernetes.io/projected/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-kube-api-access-6tlmh\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.415291 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.720911 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-7khps"] Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.772685 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-metrics-certs\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.772773 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-memberlist\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: E1124 07:00:44.772956 4799 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 07:00:44 crc kubenswrapper[4799]: E1124 07:00:44.773033 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-memberlist podName:8a7ba0ee-de14-43e5-8b93-9b915e8d9938 nodeName:}" failed. No retries permitted until 2025-11-24 07:00:45.773012183 +0000 UTC m=+791.428994657 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-memberlist") pod "speaker-bzp2s" (UID: "8a7ba0ee-de14-43e5-8b93-9b915e8d9938") : secret "metallb-memberlist" not found Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.780705 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-metrics-certs\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:44 crc kubenswrapper[4799]: I1124 07:00:44.823032 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-gkkl2"] Nov 24 07:00:45 crc kubenswrapper[4799]: I1124 07:00:45.397730 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" event={"ID":"6cd58852-5a3f-46aa-b43a-6a8ab49711a5","Type":"ContainerStarted","Data":"2f59ba9e083c96f3bc932f3597ae2de09271762f2a26b1defa2a5035a0b7a887"} Nov 24 07:00:45 crc kubenswrapper[4799]: I1124 07:00:45.400461 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2gqq5" event={"ID":"aa96ad1b-f53e-428e-ac55-bb15c9740c26","Type":"ContainerStarted","Data":"34e6b3dc764bc51f2a2c1e64067ef0c431e8f80005b62671460fa83ce9408a3b"} Nov 24 07:00:45 crc kubenswrapper[4799]: I1124 07:00:45.403212 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-gkkl2" event={"ID":"14418811-1eac-4ac6-a598-8bde1dc50a82","Type":"ContainerStarted","Data":"3fea8743ac58a68e311be358996b73f19c6a18e76cd32e7f033da633f0718f10"} Nov 24 07:00:45 crc kubenswrapper[4799]: I1124 07:00:45.403267 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-gkkl2" event={"ID":"14418811-1eac-4ac6-a598-8bde1dc50a82","Type":"ContainerStarted","Data":"c963716dddaaa45aeee294df0b967df0404cb079fff67985d196f6b4779306f1"} Nov 24 07:00:45 crc kubenswrapper[4799]: I1124 07:00:45.403288 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-gkkl2" event={"ID":"14418811-1eac-4ac6-a598-8bde1dc50a82","Type":"ContainerStarted","Data":"bcb0bc43aa436b8527c3e082a51b7ed8f76d25c89d5b4f3d8cf14d948072cf5d"} Nov 24 07:00:45 crc kubenswrapper[4799]: I1124 07:00:45.404531 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:45 crc kubenswrapper[4799]: I1124 07:00:45.427913 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-gkkl2" podStartSLOduration=1.427883357 podStartE2EDuration="1.427883357s" podCreationTimestamp="2025-11-24 07:00:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:00:45.422769502 +0000 UTC m=+791.078751996" watchObservedRunningTime="2025-11-24 07:00:45.427883357 +0000 UTC m=+791.083865851" Nov 24 07:00:45 crc kubenswrapper[4799]: I1124 07:00:45.796672 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-memberlist\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:45 crc kubenswrapper[4799]: I1124 07:00:45.801066 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8a7ba0ee-de14-43e5-8b93-9b915e8d9938-memberlist\") pod \"speaker-bzp2s\" (UID: \"8a7ba0ee-de14-43e5-8b93-9b915e8d9938\") " pod="metallb-system/speaker-bzp2s" Nov 24 07:00:45 crc kubenswrapper[4799]: I1124 07:00:45.891642 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bzp2s" Nov 24 07:00:46 crc kubenswrapper[4799]: I1124 07:00:46.411566 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bzp2s" event={"ID":"8a7ba0ee-de14-43e5-8b93-9b915e8d9938","Type":"ContainerStarted","Data":"d1a98690042f4c91f6c81fae43f985b94f71f5e7eaf9daa342c60d78a08eaf0f"} Nov 24 07:00:46 crc kubenswrapper[4799]: I1124 07:00:46.413516 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bzp2s" event={"ID":"8a7ba0ee-de14-43e5-8b93-9b915e8d9938","Type":"ContainerStarted","Data":"9a7aad02dbc74b2dadeaf1d0b6fcfd7ab70f052754e1e10537bc6e1fcd55e96f"} Nov 24 07:00:47 crc kubenswrapper[4799]: I1124 07:00:47.446158 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bzp2s" event={"ID":"8a7ba0ee-de14-43e5-8b93-9b915e8d9938","Type":"ContainerStarted","Data":"a42fea22546f7bfb04f04af6bbf020b88ec36befcb0b8e89726e71630a228664"} Nov 24 07:00:47 crc kubenswrapper[4799]: I1124 07:00:47.446745 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-bzp2s" Nov 24 07:00:47 crc kubenswrapper[4799]: I1124 07:00:47.475223 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-bzp2s" podStartSLOduration=3.475195757 podStartE2EDuration="3.475195757s" podCreationTimestamp="2025-11-24 07:00:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:00:47.472674656 +0000 UTC m=+793.128657120" watchObservedRunningTime="2025-11-24 07:00:47.475195757 +0000 UTC m=+793.131178221" Nov 24 07:00:52 crc kubenswrapper[4799]: I1124 07:00:52.482460 4799 generic.go:334] "Generic (PLEG): container finished" podID="aa96ad1b-f53e-428e-ac55-bb15c9740c26" containerID="cc965a61c0825cfa8865bc1bdeead5f3fa6d5a70a29222533f6e8368ea69f931" exitCode=0 Nov 24 07:00:52 crc kubenswrapper[4799]: I1124 07:00:52.483034 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2gqq5" event={"ID":"aa96ad1b-f53e-428e-ac55-bb15c9740c26","Type":"ContainerDied","Data":"cc965a61c0825cfa8865bc1bdeead5f3fa6d5a70a29222533f6e8368ea69f931"} Nov 24 07:00:52 crc kubenswrapper[4799]: I1124 07:00:52.485470 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" event={"ID":"6cd58852-5a3f-46aa-b43a-6a8ab49711a5","Type":"ContainerStarted","Data":"52150ae9ea98141a480c8ab3d5fb61e20021a477e44369e8a3d5f567256d7320"} Nov 24 07:00:52 crc kubenswrapper[4799]: I1124 07:00:52.485686 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" Nov 24 07:00:52 crc kubenswrapper[4799]: I1124 07:00:52.538604 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" podStartSLOduration=2.7116571929999997 podStartE2EDuration="9.538573318s" podCreationTimestamp="2025-11-24 07:00:43 +0000 UTC" firstStartedPulling="2025-11-24 07:00:44.72836913 +0000 UTC m=+790.384351604" lastFinishedPulling="2025-11-24 07:00:51.555285235 +0000 UTC m=+797.211267729" observedRunningTime="2025-11-24 07:00:52.53439057 +0000 UTC m=+798.190373054" watchObservedRunningTime="2025-11-24 07:00:52.538573318 +0000 UTC m=+798.194555822" Nov 24 07:00:53 crc kubenswrapper[4799]: I1124 07:00:53.495183 4799 generic.go:334] "Generic (PLEG): container finished" podID="aa96ad1b-f53e-428e-ac55-bb15c9740c26" containerID="1a15ce3a2727d83325d7a37170f4d588af5632ff79afb937621d2aa4af110a82" exitCode=0 Nov 24 07:00:53 crc kubenswrapper[4799]: I1124 07:00:53.495399 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2gqq5" event={"ID":"aa96ad1b-f53e-428e-ac55-bb15c9740c26","Type":"ContainerDied","Data":"1a15ce3a2727d83325d7a37170f4d588af5632ff79afb937621d2aa4af110a82"} Nov 24 07:00:54 crc kubenswrapper[4799]: I1124 07:00:54.419992 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-gkkl2" Nov 24 07:00:54 crc kubenswrapper[4799]: I1124 07:00:54.504511 4799 generic.go:334] "Generic (PLEG): container finished" podID="aa96ad1b-f53e-428e-ac55-bb15c9740c26" containerID="4ef2a2cb708686be9feb4579f5c098e5e4b6cdaa0d51b7c850596e16dd66afd2" exitCode=0 Nov 24 07:00:54 crc kubenswrapper[4799]: I1124 07:00:54.504583 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2gqq5" event={"ID":"aa96ad1b-f53e-428e-ac55-bb15c9740c26","Type":"ContainerDied","Data":"4ef2a2cb708686be9feb4579f5c098e5e4b6cdaa0d51b7c850596e16dd66afd2"} Nov 24 07:00:55 crc kubenswrapper[4799]: I1124 07:00:55.521510 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2gqq5" event={"ID":"aa96ad1b-f53e-428e-ac55-bb15c9740c26","Type":"ContainerStarted","Data":"2447b267986f9666647dda102b41ae0a2eecb6cd4419bee481771a4ee92bb3ae"} Nov 24 07:00:55 crc kubenswrapper[4799]: I1124 07:00:55.522738 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2gqq5" event={"ID":"aa96ad1b-f53e-428e-ac55-bb15c9740c26","Type":"ContainerStarted","Data":"982908df33014c23b2fe729bd09dbfe02aa1f9b1a0e95ba21910d169f274c7fe"} Nov 24 07:00:55 crc kubenswrapper[4799]: I1124 07:00:55.522778 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2gqq5" event={"ID":"aa96ad1b-f53e-428e-ac55-bb15c9740c26","Type":"ContainerStarted","Data":"206eff625d3be1f69d4d9372ac6c629fbe2cb295415ccdb91dedf4889bee9773"} Nov 24 07:00:55 crc kubenswrapper[4799]: I1124 07:00:55.522796 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2gqq5" event={"ID":"aa96ad1b-f53e-428e-ac55-bb15c9740c26","Type":"ContainerStarted","Data":"d3d58a1cd7710b33d6fe3b8ac371debd7eedc331707f49985f0eaf68bc59c2aa"} Nov 24 07:00:55 crc kubenswrapper[4799]: I1124 07:00:55.522814 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2gqq5" event={"ID":"aa96ad1b-f53e-428e-ac55-bb15c9740c26","Type":"ContainerStarted","Data":"8f3683edddd94c580dbd16c97ef7be08996343961e1dbd55a2848bff1957ddbf"} Nov 24 07:00:56 crc kubenswrapper[4799]: I1124 07:00:56.544695 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2gqq5" event={"ID":"aa96ad1b-f53e-428e-ac55-bb15c9740c26","Type":"ContainerStarted","Data":"f558751dcd428f417d34097f60d0d216907b7854053e3bbd2a2becb788104a38"} Nov 24 07:00:56 crc kubenswrapper[4799]: I1124 07:00:56.546038 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:56 crc kubenswrapper[4799]: I1124 07:00:56.590154 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-2gqq5" podStartSLOduration=6.487708363 podStartE2EDuration="13.59012248s" podCreationTimestamp="2025-11-24 07:00:43 +0000 UTC" firstStartedPulling="2025-11-24 07:00:44.448893355 +0000 UTC m=+790.104875829" lastFinishedPulling="2025-11-24 07:00:51.551307472 +0000 UTC m=+797.207289946" observedRunningTime="2025-11-24 07:00:56.582125624 +0000 UTC m=+802.238108148" watchObservedRunningTime="2025-11-24 07:00:56.59012248 +0000 UTC m=+802.246104964" Nov 24 07:00:59 crc kubenswrapper[4799]: I1124 07:00:59.292692 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:00:59 crc kubenswrapper[4799]: I1124 07:00:59.359652 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.600267 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xfwst"] Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.601794 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.617485 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfwst"] Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.674746 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-utilities\") pod \"redhat-marketplace-xfwst\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.674841 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f57wb\" (UniqueName: \"kubernetes.io/projected/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-kube-api-access-f57wb\") pod \"redhat-marketplace-xfwst\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.674878 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-catalog-content\") pod \"redhat-marketplace-xfwst\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.776575 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-utilities\") pod \"redhat-marketplace-xfwst\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.776687 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f57wb\" (UniqueName: \"kubernetes.io/projected/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-kube-api-access-f57wb\") pod \"redhat-marketplace-xfwst\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.776722 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-catalog-content\") pod \"redhat-marketplace-xfwst\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.777184 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-utilities\") pod \"redhat-marketplace-xfwst\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.777353 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-catalog-content\") pod \"redhat-marketplace-xfwst\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.811973 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f57wb\" (UniqueName: \"kubernetes.io/projected/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-kube-api-access-f57wb\") pod \"redhat-marketplace-xfwst\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:01 crc kubenswrapper[4799]: I1124 07:01:01.926096 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:02 crc kubenswrapper[4799]: I1124 07:01:02.412506 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfwst"] Nov 24 07:01:02 crc kubenswrapper[4799]: I1124 07:01:02.598056 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfwst" event={"ID":"a7ac5bd1-87af-4152-8249-53f0a1bf70cd","Type":"ContainerStarted","Data":"601579687da5322a4f29c685a863a59701c0943649a41bb29697b3f2221dad1e"} Nov 24 07:01:03 crc kubenswrapper[4799]: I1124 07:01:03.641260 4799 generic.go:334] "Generic (PLEG): container finished" podID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" containerID="f75e402e3e8169fbe7bd14c2bb9a42d875e3c5d8e2992bc7092f643e108699e2" exitCode=0 Nov 24 07:01:03 crc kubenswrapper[4799]: I1124 07:01:03.643209 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfwst" event={"ID":"a7ac5bd1-87af-4152-8249-53f0a1bf70cd","Type":"ContainerDied","Data":"f75e402e3e8169fbe7bd14c2bb9a42d875e3c5d8e2992bc7092f643e108699e2"} Nov 24 07:01:04 crc kubenswrapper[4799]: I1124 07:01:04.296185 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-2gqq5" Nov 24 07:01:04 crc kubenswrapper[4799]: I1124 07:01:04.297128 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-7khps" Nov 24 07:01:04 crc kubenswrapper[4799]: I1124 07:01:04.654021 4799 generic.go:334] "Generic (PLEG): container finished" podID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" containerID="a18a7058846472003b1509988d91d3fd9e517ff56b56d5a86605bf5345cb956a" exitCode=0 Nov 24 07:01:04 crc kubenswrapper[4799]: I1124 07:01:04.654192 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfwst" event={"ID":"a7ac5bd1-87af-4152-8249-53f0a1bf70cd","Type":"ContainerDied","Data":"a18a7058846472003b1509988d91d3fd9e517ff56b56d5a86605bf5345cb956a"} Nov 24 07:01:05 crc kubenswrapper[4799]: I1124 07:01:05.665314 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfwst" event={"ID":"a7ac5bd1-87af-4152-8249-53f0a1bf70cd","Type":"ContainerStarted","Data":"326b14ddbe7827343540d870f2f9d43691e9c3aa2aa77b2d9a90d573ad2be116"} Nov 24 07:01:05 crc kubenswrapper[4799]: I1124 07:01:05.698971 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xfwst" podStartSLOduration=3.276507656 podStartE2EDuration="4.698948711s" podCreationTimestamp="2025-11-24 07:01:01 +0000 UTC" firstStartedPulling="2025-11-24 07:01:03.644116419 +0000 UTC m=+809.300098903" lastFinishedPulling="2025-11-24 07:01:05.066557444 +0000 UTC m=+810.722539958" observedRunningTime="2025-11-24 07:01:05.69610334 +0000 UTC m=+811.352085854" watchObservedRunningTime="2025-11-24 07:01:05.698948711 +0000 UTC m=+811.354931195" Nov 24 07:01:05 crc kubenswrapper[4799]: I1124 07:01:05.896214 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-bzp2s" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.648262 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb"] Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.650489 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.656900 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.678236 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb"] Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.801246 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.801396 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxsts\" (UniqueName: \"kubernetes.io/projected/df363f0a-38fa-4510-9a1a-947e628b5f89-kube-api-access-rxsts\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.801419 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.903000 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.903108 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.903187 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxsts\" (UniqueName: \"kubernetes.io/projected/df363f0a-38fa-4510-9a1a-947e628b5f89-kube-api-access-rxsts\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.903714 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.903703 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.924331 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxsts\" (UniqueName: \"kubernetes.io/projected/df363f0a-38fa-4510-9a1a-947e628b5f89-kube-api-access-rxsts\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:07 crc kubenswrapper[4799]: I1124 07:01:07.977188 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:08 crc kubenswrapper[4799]: I1124 07:01:08.208276 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb"] Nov 24 07:01:08 crc kubenswrapper[4799]: I1124 07:01:08.690262 4799 generic.go:334] "Generic (PLEG): container finished" podID="df363f0a-38fa-4510-9a1a-947e628b5f89" containerID="7e43ef193e93c7861f1e1e79ba783e2f79bb3f7950cdad4f48eafefd993e05ca" exitCode=0 Nov 24 07:01:08 crc kubenswrapper[4799]: I1124 07:01:08.690383 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" event={"ID":"df363f0a-38fa-4510-9a1a-947e628b5f89","Type":"ContainerDied","Data":"7e43ef193e93c7861f1e1e79ba783e2f79bb3f7950cdad4f48eafefd993e05ca"} Nov 24 07:01:08 crc kubenswrapper[4799]: I1124 07:01:08.690828 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" event={"ID":"df363f0a-38fa-4510-9a1a-947e628b5f89","Type":"ContainerStarted","Data":"195a0400abf2d100d5eb08f5927978f9bb7fe8852f4282e6ae480faad921ff5d"} Nov 24 07:01:08 crc kubenswrapper[4799]: I1124 07:01:08.988070 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pdckk"] Nov 24 07:01:08 crc kubenswrapper[4799]: I1124 07:01:08.989314 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.019351 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pdckk"] Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.123009 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cb92\" (UniqueName: \"kubernetes.io/projected/5935c614-4a8b-485f-84ff-a84cb8fe26df-kube-api-access-2cb92\") pod \"redhat-operators-pdckk\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.123098 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-catalog-content\") pod \"redhat-operators-pdckk\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.123121 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-utilities\") pod \"redhat-operators-pdckk\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.224313 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-utilities\") pod \"redhat-operators-pdckk\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.224451 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cb92\" (UniqueName: \"kubernetes.io/projected/5935c614-4a8b-485f-84ff-a84cb8fe26df-kube-api-access-2cb92\") pod \"redhat-operators-pdckk\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.224518 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-catalog-content\") pod \"redhat-operators-pdckk\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.224992 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-utilities\") pod \"redhat-operators-pdckk\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.225093 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-catalog-content\") pod \"redhat-operators-pdckk\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.252185 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cb92\" (UniqueName: \"kubernetes.io/projected/5935c614-4a8b-485f-84ff-a84cb8fe26df-kube-api-access-2cb92\") pod \"redhat-operators-pdckk\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.342754 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:09 crc kubenswrapper[4799]: I1124 07:01:09.857549 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pdckk"] Nov 24 07:01:10 crc kubenswrapper[4799]: I1124 07:01:10.710071 4799 generic.go:334] "Generic (PLEG): container finished" podID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerID="8ad418698ba66c794f3e9ff1e964fefbbdc72d2d04e9703ceafc7864216b5d0a" exitCode=0 Nov 24 07:01:10 crc kubenswrapper[4799]: I1124 07:01:10.710171 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdckk" event={"ID":"5935c614-4a8b-485f-84ff-a84cb8fe26df","Type":"ContainerDied","Data":"8ad418698ba66c794f3e9ff1e964fefbbdc72d2d04e9703ceafc7864216b5d0a"} Nov 24 07:01:10 crc kubenswrapper[4799]: I1124 07:01:10.710533 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdckk" event={"ID":"5935c614-4a8b-485f-84ff-a84cb8fe26df","Type":"ContainerStarted","Data":"99a4e6d65134d5b3db4a774dc0fff105a36beee410754a44eed8e6982e7000bc"} Nov 24 07:01:11 crc kubenswrapper[4799]: I1124 07:01:11.928388 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:11 crc kubenswrapper[4799]: I1124 07:01:11.928615 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:11 crc kubenswrapper[4799]: I1124 07:01:11.994990 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:12 crc kubenswrapper[4799]: I1124 07:01:12.741195 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdckk" event={"ID":"5935c614-4a8b-485f-84ff-a84cb8fe26df","Type":"ContainerStarted","Data":"f09855c22119ab4dfd2a3e13acc2bbfa8edd4937414d25c597f3b8daaa68f7e7"} Nov 24 07:01:12 crc kubenswrapper[4799]: I1124 07:01:12.744352 4799 generic.go:334] "Generic (PLEG): container finished" podID="df363f0a-38fa-4510-9a1a-947e628b5f89" containerID="202a1acb850f66ae554db70c96c48df93db4f54f8fef2648d703a87bd44b9429" exitCode=0 Nov 24 07:01:12 crc kubenswrapper[4799]: I1124 07:01:12.745347 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" event={"ID":"df363f0a-38fa-4510-9a1a-947e628b5f89","Type":"ContainerDied","Data":"202a1acb850f66ae554db70c96c48df93db4f54f8fef2648d703a87bd44b9429"} Nov 24 07:01:12 crc kubenswrapper[4799]: I1124 07:01:12.798673 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:13 crc kubenswrapper[4799]: I1124 07:01:13.755882 4799 generic.go:334] "Generic (PLEG): container finished" podID="df363f0a-38fa-4510-9a1a-947e628b5f89" containerID="192b3c2596f32ce9d0fb7e02e08a002a6522af97351e864346b12d50ce964fe9" exitCode=0 Nov 24 07:01:13 crc kubenswrapper[4799]: I1124 07:01:13.755989 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" event={"ID":"df363f0a-38fa-4510-9a1a-947e628b5f89","Type":"ContainerDied","Data":"192b3c2596f32ce9d0fb7e02e08a002a6522af97351e864346b12d50ce964fe9"} Nov 24 07:01:13 crc kubenswrapper[4799]: I1124 07:01:13.758979 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdckk" event={"ID":"5935c614-4a8b-485f-84ff-a84cb8fe26df","Type":"ContainerDied","Data":"f09855c22119ab4dfd2a3e13acc2bbfa8edd4937414d25c597f3b8daaa68f7e7"} Nov 24 07:01:13 crc kubenswrapper[4799]: I1124 07:01:13.758941 4799 generic.go:334] "Generic (PLEG): container finished" podID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerID="f09855c22119ab4dfd2a3e13acc2bbfa8edd4937414d25c597f3b8daaa68f7e7" exitCode=0 Nov 24 07:01:14 crc kubenswrapper[4799]: I1124 07:01:14.769625 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdckk" event={"ID":"5935c614-4a8b-485f-84ff-a84cb8fe26df","Type":"ContainerStarted","Data":"9ae64cf7b3a06b7a2835ebfd136c518b084d9dc1fa6b85d9cd58b61db13bfce6"} Nov 24 07:01:14 crc kubenswrapper[4799]: I1124 07:01:14.799369 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pdckk" podStartSLOduration=4.214913361 podStartE2EDuration="6.799341473s" podCreationTimestamp="2025-11-24 07:01:08 +0000 UTC" firstStartedPulling="2025-11-24 07:01:11.789196189 +0000 UTC m=+817.445178663" lastFinishedPulling="2025-11-24 07:01:14.373624281 +0000 UTC m=+820.029606775" observedRunningTime="2025-11-24 07:01:14.796969896 +0000 UTC m=+820.452952400" watchObservedRunningTime="2025-11-24 07:01:14.799341473 +0000 UTC m=+820.455323947" Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.098648 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.236892 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-bundle\") pod \"df363f0a-38fa-4510-9a1a-947e628b5f89\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.237421 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-util\") pod \"df363f0a-38fa-4510-9a1a-947e628b5f89\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.237540 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxsts\" (UniqueName: \"kubernetes.io/projected/df363f0a-38fa-4510-9a1a-947e628b5f89-kube-api-access-rxsts\") pod \"df363f0a-38fa-4510-9a1a-947e628b5f89\" (UID: \"df363f0a-38fa-4510-9a1a-947e628b5f89\") " Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.238892 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-bundle" (OuterVolumeSpecName: "bundle") pod "df363f0a-38fa-4510-9a1a-947e628b5f89" (UID: "df363f0a-38fa-4510-9a1a-947e628b5f89"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.248337 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-util" (OuterVolumeSpecName: "util") pod "df363f0a-38fa-4510-9a1a-947e628b5f89" (UID: "df363f0a-38fa-4510-9a1a-947e628b5f89"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.250112 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df363f0a-38fa-4510-9a1a-947e628b5f89-kube-api-access-rxsts" (OuterVolumeSpecName: "kube-api-access-rxsts") pod "df363f0a-38fa-4510-9a1a-947e628b5f89" (UID: "df363f0a-38fa-4510-9a1a-947e628b5f89"). InnerVolumeSpecName "kube-api-access-rxsts". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.339040 4799 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.339074 4799 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df363f0a-38fa-4510-9a1a-947e628b5f89-util\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.339085 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxsts\" (UniqueName: \"kubernetes.io/projected/df363f0a-38fa-4510-9a1a-947e628b5f89-kube-api-access-rxsts\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.579251 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfwst"] Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.579564 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xfwst" podUID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" containerName="registry-server" containerID="cri-o://326b14ddbe7827343540d870f2f9d43691e9c3aa2aa77b2d9a90d573ad2be116" gracePeriod=2 Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.783064 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" event={"ID":"df363f0a-38fa-4510-9a1a-947e628b5f89","Type":"ContainerDied","Data":"195a0400abf2d100d5eb08f5927978f9bb7fe8852f4282e6ae480faad921ff5d"} Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.783129 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="195a0400abf2d100d5eb08f5927978f9bb7fe8852f4282e6ae480faad921ff5d" Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.783168 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb" Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.785980 4799 generic.go:334] "Generic (PLEG): container finished" podID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" containerID="326b14ddbe7827343540d870f2f9d43691e9c3aa2aa77b2d9a90d573ad2be116" exitCode=0 Nov 24 07:01:15 crc kubenswrapper[4799]: I1124 07:01:15.786082 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfwst" event={"ID":"a7ac5bd1-87af-4152-8249-53f0a1bf70cd","Type":"ContainerDied","Data":"326b14ddbe7827343540d870f2f9d43691e9c3aa2aa77b2d9a90d573ad2be116"} Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.020609 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.073753 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-catalog-content\") pod \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.073903 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f57wb\" (UniqueName: \"kubernetes.io/projected/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-kube-api-access-f57wb\") pod \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.073948 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-utilities\") pod \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\" (UID: \"a7ac5bd1-87af-4152-8249-53f0a1bf70cd\") " Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.075510 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-utilities" (OuterVolumeSpecName: "utilities") pod "a7ac5bd1-87af-4152-8249-53f0a1bf70cd" (UID: "a7ac5bd1-87af-4152-8249-53f0a1bf70cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.083265 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-kube-api-access-f57wb" (OuterVolumeSpecName: "kube-api-access-f57wb") pod "a7ac5bd1-87af-4152-8249-53f0a1bf70cd" (UID: "a7ac5bd1-87af-4152-8249-53f0a1bf70cd"). InnerVolumeSpecName "kube-api-access-f57wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.102118 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7ac5bd1-87af-4152-8249-53f0a1bf70cd" (UID: "a7ac5bd1-87af-4152-8249-53f0a1bf70cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.175978 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.176019 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.176035 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f57wb\" (UniqueName: \"kubernetes.io/projected/a7ac5bd1-87af-4152-8249-53f0a1bf70cd-kube-api-access-f57wb\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.794296 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfwst" event={"ID":"a7ac5bd1-87af-4152-8249-53f0a1bf70cd","Type":"ContainerDied","Data":"601579687da5322a4f29c685a863a59701c0943649a41bb29697b3f2221dad1e"} Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.794789 4799 scope.go:117] "RemoveContainer" containerID="326b14ddbe7827343540d870f2f9d43691e9c3aa2aa77b2d9a90d573ad2be116" Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.794457 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfwst" Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.819240 4799 scope.go:117] "RemoveContainer" containerID="a18a7058846472003b1509988d91d3fd9e517ff56b56d5a86605bf5345cb956a" Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.836155 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfwst"] Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.846892 4799 scope.go:117] "RemoveContainer" containerID="f75e402e3e8169fbe7bd14c2bb9a42d875e3c5d8e2992bc7092f643e108699e2" Nov 24 07:01:16 crc kubenswrapper[4799]: I1124 07:01:16.847217 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfwst"] Nov 24 07:01:16 crc kubenswrapper[4799]: E1124 07:01:16.922146 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7ac5bd1_87af_4152_8249_53f0a1bf70cd.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7ac5bd1_87af_4152_8249_53f0a1bf70cd.slice/crio-601579687da5322a4f29c685a863a59701c0943649a41bb29697b3f2221dad1e\": RecentStats: unable to find data in memory cache]" Nov 24 07:01:17 crc kubenswrapper[4799]: I1124 07:01:17.640758 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" path="/var/lib/kubelet/pods/a7ac5bd1-87af-4152-8249-53f0a1bf70cd/volumes" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.311054 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv"] Nov 24 07:01:19 crc kubenswrapper[4799]: E1124 07:01:19.311898 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df363f0a-38fa-4510-9a1a-947e628b5f89" containerName="util" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.311918 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="df363f0a-38fa-4510-9a1a-947e628b5f89" containerName="util" Nov 24 07:01:19 crc kubenswrapper[4799]: E1124 07:01:19.311933 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" containerName="registry-server" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.311942 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" containerName="registry-server" Nov 24 07:01:19 crc kubenswrapper[4799]: E1124 07:01:19.311963 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df363f0a-38fa-4510-9a1a-947e628b5f89" containerName="pull" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.311970 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="df363f0a-38fa-4510-9a1a-947e628b5f89" containerName="pull" Nov 24 07:01:19 crc kubenswrapper[4799]: E1124 07:01:19.311981 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" containerName="extract-utilities" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.311989 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" containerName="extract-utilities" Nov 24 07:01:19 crc kubenswrapper[4799]: E1124 07:01:19.312006 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df363f0a-38fa-4510-9a1a-947e628b5f89" containerName="extract" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.312014 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="df363f0a-38fa-4510-9a1a-947e628b5f89" containerName="extract" Nov 24 07:01:19 crc kubenswrapper[4799]: E1124 07:01:19.312029 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" containerName="extract-content" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.312036 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" containerName="extract-content" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.312172 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ac5bd1-87af-4152-8249-53f0a1bf70cd" containerName="registry-server" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.312196 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="df363f0a-38fa-4510-9a1a-947e628b5f89" containerName="extract" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.312737 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.316259 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.316513 4799 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-7ds7f" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.316685 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.332224 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv"] Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.343317 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.343395 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.421320 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/cc7672ac-cc75-4e0c-be06-ee1696dee935-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-sh5rv\" (UID: \"cc7672ac-cc75-4e0c-be06-ee1696dee935\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.421399 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk2bw\" (UniqueName: \"kubernetes.io/projected/cc7672ac-cc75-4e0c-be06-ee1696dee935-kube-api-access-fk2bw\") pod \"cert-manager-operator-controller-manager-64cf6dff88-sh5rv\" (UID: \"cc7672ac-cc75-4e0c-be06-ee1696dee935\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.523644 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/cc7672ac-cc75-4e0c-be06-ee1696dee935-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-sh5rv\" (UID: \"cc7672ac-cc75-4e0c-be06-ee1696dee935\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.523715 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk2bw\" (UniqueName: \"kubernetes.io/projected/cc7672ac-cc75-4e0c-be06-ee1696dee935-kube-api-access-fk2bw\") pod \"cert-manager-operator-controller-manager-64cf6dff88-sh5rv\" (UID: \"cc7672ac-cc75-4e0c-be06-ee1696dee935\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.524383 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/cc7672ac-cc75-4e0c-be06-ee1696dee935-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-sh5rv\" (UID: \"cc7672ac-cc75-4e0c-be06-ee1696dee935\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.548652 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk2bw\" (UniqueName: \"kubernetes.io/projected/cc7672ac-cc75-4e0c-be06-ee1696dee935-kube-api-access-fk2bw\") pod \"cert-manager-operator-controller-manager-64cf6dff88-sh5rv\" (UID: \"cc7672ac-cc75-4e0c-be06-ee1696dee935\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.634203 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv" Nov 24 07:01:19 crc kubenswrapper[4799]: I1124 07:01:19.991926 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv"] Nov 24 07:01:20 crc kubenswrapper[4799]: I1124 07:01:20.391079 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pdckk" podUID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerName="registry-server" probeResult="failure" output=< Nov 24 07:01:20 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 07:01:20 crc kubenswrapper[4799]: > Nov 24 07:01:20 crc kubenswrapper[4799]: I1124 07:01:20.839598 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv" event={"ID":"cc7672ac-cc75-4e0c-be06-ee1696dee935","Type":"ContainerStarted","Data":"c2637a31943a22b363b10eb702a43d520c9d6f3865a0b14c2657525b2a5b4bbc"} Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.777033 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kcqk4"] Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.779561 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.798521 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kcqk4"] Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.862397 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-catalog-content\") pod \"community-operators-kcqk4\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.862719 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-utilities\") pod \"community-operators-kcqk4\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.862824 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htjnv\" (UniqueName: \"kubernetes.io/projected/e52f5c42-2d4e-4d15-a277-1e8391f69e00-kube-api-access-htjnv\") pod \"community-operators-kcqk4\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.963999 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-catalog-content\") pod \"community-operators-kcqk4\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.964079 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-utilities\") pod \"community-operators-kcqk4\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.964125 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htjnv\" (UniqueName: \"kubernetes.io/projected/e52f5c42-2d4e-4d15-a277-1e8391f69e00-kube-api-access-htjnv\") pod \"community-operators-kcqk4\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.965023 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-utilities\") pod \"community-operators-kcqk4\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.965101 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-catalog-content\") pod \"community-operators-kcqk4\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:21 crc kubenswrapper[4799]: I1124 07:01:21.995228 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htjnv\" (UniqueName: \"kubernetes.io/projected/e52f5c42-2d4e-4d15-a277-1e8391f69e00-kube-api-access-htjnv\") pod \"community-operators-kcqk4\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:22 crc kubenswrapper[4799]: I1124 07:01:22.110014 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:22 crc kubenswrapper[4799]: I1124 07:01:22.432767 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kcqk4"] Nov 24 07:01:22 crc kubenswrapper[4799]: I1124 07:01:22.861651 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kcqk4" event={"ID":"e52f5c42-2d4e-4d15-a277-1e8391f69e00","Type":"ContainerStarted","Data":"49251d28eda67ff0c26bbb49963fd4ce75d8133129e17939307209c44883b71d"} Nov 24 07:01:23 crc kubenswrapper[4799]: I1124 07:01:23.875429 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kcqk4" event={"ID":"e52f5c42-2d4e-4d15-a277-1e8391f69e00","Type":"ContainerStarted","Data":"fb51fda68c12afdd26d507e56f5943c656540767b245f95ae678e206d8b591be"} Nov 24 07:01:24 crc kubenswrapper[4799]: I1124 07:01:24.885387 4799 generic.go:334] "Generic (PLEG): container finished" podID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" containerID="fb51fda68c12afdd26d507e56f5943c656540767b245f95ae678e206d8b591be" exitCode=0 Nov 24 07:01:24 crc kubenswrapper[4799]: I1124 07:01:24.885495 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kcqk4" event={"ID":"e52f5c42-2d4e-4d15-a277-1e8391f69e00","Type":"ContainerDied","Data":"fb51fda68c12afdd26d507e56f5943c656540767b245f95ae678e206d8b591be"} Nov 24 07:01:27 crc kubenswrapper[4799]: I1124 07:01:27.941674 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv" event={"ID":"cc7672ac-cc75-4e0c-be06-ee1696dee935","Type":"ContainerStarted","Data":"34828ecb9f72df668d5a77d9dea8c77cae4b1b5a4acf5067decc6b5c680ee65f"} Nov 24 07:01:27 crc kubenswrapper[4799]: I1124 07:01:27.944372 4799 generic.go:334] "Generic (PLEG): container finished" podID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" containerID="6050077eafe16fadc43066842743bfb05702c8fe964e70eb3f1c364f9fe56970" exitCode=0 Nov 24 07:01:27 crc kubenswrapper[4799]: I1124 07:01:27.944444 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kcqk4" event={"ID":"e52f5c42-2d4e-4d15-a277-1e8391f69e00","Type":"ContainerDied","Data":"6050077eafe16fadc43066842743bfb05702c8fe964e70eb3f1c364f9fe56970"} Nov 24 07:01:27 crc kubenswrapper[4799]: I1124 07:01:27.965882 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-sh5rv" podStartSLOduration=1.722674718 podStartE2EDuration="8.965819776s" podCreationTimestamp="2025-11-24 07:01:19 +0000 UTC" firstStartedPulling="2025-11-24 07:01:20.012951083 +0000 UTC m=+825.668933557" lastFinishedPulling="2025-11-24 07:01:27.256096151 +0000 UTC m=+832.912078615" observedRunningTime="2025-11-24 07:01:27.962974876 +0000 UTC m=+833.618957360" watchObservedRunningTime="2025-11-24 07:01:27.965819776 +0000 UTC m=+833.621802250" Nov 24 07:01:27 crc kubenswrapper[4799]: I1124 07:01:27.990067 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x5t8r"] Nov 24 07:01:27 crc kubenswrapper[4799]: I1124 07:01:27.991515 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.016026 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x5t8r"] Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.158952 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-utilities\") pod \"certified-operators-x5t8r\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.159025 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpq4z\" (UniqueName: \"kubernetes.io/projected/4e2791a7-9b30-427a-872e-cdb4967dc4b3-kube-api-access-mpq4z\") pod \"certified-operators-x5t8r\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.159150 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-catalog-content\") pod \"certified-operators-x5t8r\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.260273 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-catalog-content\") pod \"certified-operators-x5t8r\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.260378 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-utilities\") pod \"certified-operators-x5t8r\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.260405 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpq4z\" (UniqueName: \"kubernetes.io/projected/4e2791a7-9b30-427a-872e-cdb4967dc4b3-kube-api-access-mpq4z\") pod \"certified-operators-x5t8r\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.260812 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-catalog-content\") pod \"certified-operators-x5t8r\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.261084 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-utilities\") pod \"certified-operators-x5t8r\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.284978 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpq4z\" (UniqueName: \"kubernetes.io/projected/4e2791a7-9b30-427a-872e-cdb4967dc4b3-kube-api-access-mpq4z\") pod \"certified-operators-x5t8r\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.306600 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.727298 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x5t8r"] Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.958596 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kcqk4" event={"ID":"e52f5c42-2d4e-4d15-a277-1e8391f69e00","Type":"ContainerStarted","Data":"a7b0dcc8df9b07f16c6ce95454b7d789f79b2cbcea3329b2a02f7309c7e9b69f"} Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.961182 4799 generic.go:334] "Generic (PLEG): container finished" podID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" containerID="0a0eafdeb538f21ab5c1a57460a2d358a1e67beef51090b7748071efc6e8c36d" exitCode=0 Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.961354 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5t8r" event={"ID":"4e2791a7-9b30-427a-872e-cdb4967dc4b3","Type":"ContainerDied","Data":"0a0eafdeb538f21ab5c1a57460a2d358a1e67beef51090b7748071efc6e8c36d"} Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.961432 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5t8r" event={"ID":"4e2791a7-9b30-427a-872e-cdb4967dc4b3","Type":"ContainerStarted","Data":"c28662a6b672c0c1f8b9578a3d5797532e6d43e381d1e0f6ea0e15c8a78f8799"} Nov 24 07:01:28 crc kubenswrapper[4799]: I1124 07:01:28.984109 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kcqk4" podStartSLOduration=5.809201531 podStartE2EDuration="7.984081479s" podCreationTimestamp="2025-11-24 07:01:21 +0000 UTC" firstStartedPulling="2025-11-24 07:01:26.152099874 +0000 UTC m=+831.808082358" lastFinishedPulling="2025-11-24 07:01:28.326979832 +0000 UTC m=+833.982962306" observedRunningTime="2025-11-24 07:01:28.979576901 +0000 UTC m=+834.635559375" watchObservedRunningTime="2025-11-24 07:01:28.984081479 +0000 UTC m=+834.640063953" Nov 24 07:01:29 crc kubenswrapper[4799]: I1124 07:01:29.387936 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:29 crc kubenswrapper[4799]: I1124 07:01:29.445498 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:29 crc kubenswrapper[4799]: I1124 07:01:29.969816 4799 generic.go:334] "Generic (PLEG): container finished" podID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" containerID="1b3119925156ae14974e027ca914887526567e64f4b5e44e0730e5021283682c" exitCode=0 Nov 24 07:01:29 crc kubenswrapper[4799]: I1124 07:01:29.969938 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5t8r" event={"ID":"4e2791a7-9b30-427a-872e-cdb4967dc4b3","Type":"ContainerDied","Data":"1b3119925156ae14974e027ca914887526567e64f4b5e44e0730e5021283682c"} Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.290693 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-58pnr"] Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.291705 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.294920 4799 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-69xdl" Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.296493 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.298161 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.312220 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-58pnr"] Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.392990 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a45abe66-dc32-4b76-90d9-b821dba28bd5-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-58pnr\" (UID: \"a45abe66-dc32-4b76-90d9-b821dba28bd5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.393099 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrw7j\" (UniqueName: \"kubernetes.io/projected/a45abe66-dc32-4b76-90d9-b821dba28bd5-kube-api-access-lrw7j\") pod \"cert-manager-webhook-f4fb5df64-58pnr\" (UID: \"a45abe66-dc32-4b76-90d9-b821dba28bd5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.493939 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a45abe66-dc32-4b76-90d9-b821dba28bd5-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-58pnr\" (UID: \"a45abe66-dc32-4b76-90d9-b821dba28bd5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.494336 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrw7j\" (UniqueName: \"kubernetes.io/projected/a45abe66-dc32-4b76-90d9-b821dba28bd5-kube-api-access-lrw7j\") pod \"cert-manager-webhook-f4fb5df64-58pnr\" (UID: \"a45abe66-dc32-4b76-90d9-b821dba28bd5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.525700 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a45abe66-dc32-4b76-90d9-b821dba28bd5-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-58pnr\" (UID: \"a45abe66-dc32-4b76-90d9-b821dba28bd5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.525923 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrw7j\" (UniqueName: \"kubernetes.io/projected/a45abe66-dc32-4b76-90d9-b821dba28bd5-kube-api-access-lrw7j\") pod \"cert-manager-webhook-f4fb5df64-58pnr\" (UID: \"a45abe66-dc32-4b76-90d9-b821dba28bd5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.608460 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" Nov 24 07:01:30 crc kubenswrapper[4799]: I1124 07:01:30.980270 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5t8r" event={"ID":"4e2791a7-9b30-427a-872e-cdb4967dc4b3","Type":"ContainerStarted","Data":"ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7"} Nov 24 07:01:31 crc kubenswrapper[4799]: I1124 07:01:31.003796 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x5t8r" podStartSLOduration=2.595062301 podStartE2EDuration="4.003773938s" podCreationTimestamp="2025-11-24 07:01:27 +0000 UTC" firstStartedPulling="2025-11-24 07:01:28.964489265 +0000 UTC m=+834.620471739" lastFinishedPulling="2025-11-24 07:01:30.373200902 +0000 UTC m=+836.029183376" observedRunningTime="2025-11-24 07:01:31.002701988 +0000 UTC m=+836.658684462" watchObservedRunningTime="2025-11-24 07:01:31.003773938 +0000 UTC m=+836.659756402" Nov 24 07:01:31 crc kubenswrapper[4799]: I1124 07:01:31.104725 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-58pnr"] Nov 24 07:01:32 crc kubenswrapper[4799]: I1124 07:01:32.006133 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" event={"ID":"a45abe66-dc32-4b76-90d9-b821dba28bd5","Type":"ContainerStarted","Data":"8183b1733caf7761b854656303025989830d1d1427bb0a6d319cd0fc63c6d154"} Nov 24 07:01:32 crc kubenswrapper[4799]: I1124 07:01:32.112336 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:32 crc kubenswrapper[4799]: I1124 07:01:32.112390 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:32 crc kubenswrapper[4799]: I1124 07:01:32.176721 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:33 crc kubenswrapper[4799]: I1124 07:01:33.059698 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:33 crc kubenswrapper[4799]: I1124 07:01:33.174883 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pdckk"] Nov 24 07:01:33 crc kubenswrapper[4799]: I1124 07:01:33.175329 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pdckk" podUID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerName="registry-server" containerID="cri-o://9ae64cf7b3a06b7a2835ebfd136c518b084d9dc1fa6b85d9cd58b61db13bfce6" gracePeriod=2 Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.363097 4799 generic.go:334] "Generic (PLEG): container finished" podID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerID="9ae64cf7b3a06b7a2835ebfd136c518b084d9dc1fa6b85d9cd58b61db13bfce6" exitCode=0 Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.363263 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdckk" event={"ID":"5935c614-4a8b-485f-84ff-a84cb8fe26df","Type":"ContainerDied","Data":"9ae64cf7b3a06b7a2835ebfd136c518b084d9dc1fa6b85d9cd58b61db13bfce6"} Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.634801 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z"] Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.635997 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z" Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.639473 4799 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-nf5t7" Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.661317 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z"] Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.738718 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtnj8\" (UniqueName: \"kubernetes.io/projected/439904d9-d1e0-4733-8f08-db0f792fdcc2-kube-api-access-mtnj8\") pod \"cert-manager-cainjector-855d9ccff4-ktx6z\" (UID: \"439904d9-d1e0-4733-8f08-db0f792fdcc2\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z" Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.738808 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/439904d9-d1e0-4733-8f08-db0f792fdcc2-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-ktx6z\" (UID: \"439904d9-d1e0-4733-8f08-db0f792fdcc2\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z" Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.770661 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kcqk4"] Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.840839 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtnj8\" (UniqueName: \"kubernetes.io/projected/439904d9-d1e0-4733-8f08-db0f792fdcc2-kube-api-access-mtnj8\") pod \"cert-manager-cainjector-855d9ccff4-ktx6z\" (UID: \"439904d9-d1e0-4733-8f08-db0f792fdcc2\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z" Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.840914 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/439904d9-d1e0-4733-8f08-db0f792fdcc2-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-ktx6z\" (UID: \"439904d9-d1e0-4733-8f08-db0f792fdcc2\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z" Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.867632 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/439904d9-d1e0-4733-8f08-db0f792fdcc2-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-ktx6z\" (UID: \"439904d9-d1e0-4733-8f08-db0f792fdcc2\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z" Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.869420 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtnj8\" (UniqueName: \"kubernetes.io/projected/439904d9-d1e0-4733-8f08-db0f792fdcc2-kube-api-access-mtnj8\") pod \"cert-manager-cainjector-855d9ccff4-ktx6z\" (UID: \"439904d9-d1e0-4733-8f08-db0f792fdcc2\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z" Nov 24 07:01:34 crc kubenswrapper[4799]: I1124 07:01:34.973535 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.054299 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.144415 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cb92\" (UniqueName: \"kubernetes.io/projected/5935c614-4a8b-485f-84ff-a84cb8fe26df-kube-api-access-2cb92\") pod \"5935c614-4a8b-485f-84ff-a84cb8fe26df\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.144475 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-catalog-content\") pod \"5935c614-4a8b-485f-84ff-a84cb8fe26df\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.144582 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-utilities\") pod \"5935c614-4a8b-485f-84ff-a84cb8fe26df\" (UID: \"5935c614-4a8b-485f-84ff-a84cb8fe26df\") " Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.146344 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-utilities" (OuterVolumeSpecName: "utilities") pod "5935c614-4a8b-485f-84ff-a84cb8fe26df" (UID: "5935c614-4a8b-485f-84ff-a84cb8fe26df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.155125 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5935c614-4a8b-485f-84ff-a84cb8fe26df-kube-api-access-2cb92" (OuterVolumeSpecName: "kube-api-access-2cb92") pod "5935c614-4a8b-485f-84ff-a84cb8fe26df" (UID: "5935c614-4a8b-485f-84ff-a84cb8fe26df"). InnerVolumeSpecName "kube-api-access-2cb92". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.245956 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.245998 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cb92\" (UniqueName: \"kubernetes.io/projected/5935c614-4a8b-485f-84ff-a84cb8fe26df-kube-api-access-2cb92\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.316133 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5935c614-4a8b-485f-84ff-a84cb8fe26df" (UID: "5935c614-4a8b-485f-84ff-a84cb8fe26df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.347078 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5935c614-4a8b-485f-84ff-a84cb8fe26df-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.379958 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdckk" event={"ID":"5935c614-4a8b-485f-84ff-a84cb8fe26df","Type":"ContainerDied","Data":"99a4e6d65134d5b3db4a774dc0fff105a36beee410754a44eed8e6982e7000bc"} Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.380013 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdckk" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.380062 4799 scope.go:117] "RemoveContainer" containerID="9ae64cf7b3a06b7a2835ebfd136c518b084d9dc1fa6b85d9cd58b61db13bfce6" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.380156 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kcqk4" podUID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" containerName="registry-server" containerID="cri-o://a7b0dcc8df9b07f16c6ce95454b7d789f79b2cbcea3329b2a02f7309c7e9b69f" gracePeriod=2 Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.412131 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pdckk"] Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.419869 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pdckk"] Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.422868 4799 scope.go:117] "RemoveContainer" containerID="f09855c22119ab4dfd2a3e13acc2bbfa8edd4937414d25c597f3b8daaa68f7e7" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.454129 4799 scope.go:117] "RemoveContainer" containerID="8ad418698ba66c794f3e9ff1e964fefbbdc72d2d04e9703ceafc7864216b5d0a" Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.588580 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z"] Nov 24 07:01:35 crc kubenswrapper[4799]: W1124 07:01:35.642622 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod439904d9_d1e0_4733_8f08_db0f792fdcc2.slice/crio-c3d8819029a61dd2920e3c51f3d03b68143d363aacba35ea013a1f252eb380c3 WatchSource:0}: Error finding container c3d8819029a61dd2920e3c51f3d03b68143d363aacba35ea013a1f252eb380c3: Status 404 returned error can't find the container with id c3d8819029a61dd2920e3c51f3d03b68143d363aacba35ea013a1f252eb380c3 Nov 24 07:01:35 crc kubenswrapper[4799]: I1124 07:01:35.659518 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5935c614-4a8b-485f-84ff-a84cb8fe26df" path="/var/lib/kubelet/pods/5935c614-4a8b-485f-84ff-a84cb8fe26df/volumes" Nov 24 07:01:36 crc kubenswrapper[4799]: I1124 07:01:36.391109 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z" event={"ID":"439904d9-d1e0-4733-8f08-db0f792fdcc2","Type":"ContainerStarted","Data":"c3d8819029a61dd2920e3c51f3d03b68143d363aacba35ea013a1f252eb380c3"} Nov 24 07:01:36 crc kubenswrapper[4799]: I1124 07:01:36.395582 4799 generic.go:334] "Generic (PLEG): container finished" podID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" containerID="a7b0dcc8df9b07f16c6ce95454b7d789f79b2cbcea3329b2a02f7309c7e9b69f" exitCode=0 Nov 24 07:01:36 crc kubenswrapper[4799]: I1124 07:01:36.395740 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kcqk4" event={"ID":"e52f5c42-2d4e-4d15-a277-1e8391f69e00","Type":"ContainerDied","Data":"a7b0dcc8df9b07f16c6ce95454b7d789f79b2cbcea3329b2a02f7309c7e9b69f"} Nov 24 07:01:38 crc kubenswrapper[4799]: I1124 07:01:38.307730 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:38 crc kubenswrapper[4799]: I1124 07:01:38.308281 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:38 crc kubenswrapper[4799]: I1124 07:01:38.361749 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:38 crc kubenswrapper[4799]: I1124 07:01:38.453321 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:38 crc kubenswrapper[4799]: I1124 07:01:38.961415 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.006857 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-utilities\") pod \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.007006 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htjnv\" (UniqueName: \"kubernetes.io/projected/e52f5c42-2d4e-4d15-a277-1e8391f69e00-kube-api-access-htjnv\") pod \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.007784 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-utilities" (OuterVolumeSpecName: "utilities") pod "e52f5c42-2d4e-4d15-a277-1e8391f69e00" (UID: "e52f5c42-2d4e-4d15-a277-1e8391f69e00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.008002 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-catalog-content\") pod \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\" (UID: \"e52f5c42-2d4e-4d15-a277-1e8391f69e00\") " Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.008648 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.014777 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e52f5c42-2d4e-4d15-a277-1e8391f69e00-kube-api-access-htjnv" (OuterVolumeSpecName: "kube-api-access-htjnv") pod "e52f5c42-2d4e-4d15-a277-1e8391f69e00" (UID: "e52f5c42-2d4e-4d15-a277-1e8391f69e00"). InnerVolumeSpecName "kube-api-access-htjnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.059554 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e52f5c42-2d4e-4d15-a277-1e8391f69e00" (UID: "e52f5c42-2d4e-4d15-a277-1e8391f69e00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.110229 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htjnv\" (UniqueName: \"kubernetes.io/projected/e52f5c42-2d4e-4d15-a277-1e8391f69e00-kube-api-access-htjnv\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.110276 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52f5c42-2d4e-4d15-a277-1e8391f69e00-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.426169 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kcqk4" Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.426240 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kcqk4" event={"ID":"e52f5c42-2d4e-4d15-a277-1e8391f69e00","Type":"ContainerDied","Data":"49251d28eda67ff0c26bbb49963fd4ce75d8133129e17939307209c44883b71d"} Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.426300 4799 scope.go:117] "RemoveContainer" containerID="a7b0dcc8df9b07f16c6ce95454b7d789f79b2cbcea3329b2a02f7309c7e9b69f" Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.475907 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kcqk4"] Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.480093 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kcqk4"] Nov 24 07:01:39 crc kubenswrapper[4799]: I1124 07:01:39.637698 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" path="/var/lib/kubelet/pods/e52f5c42-2d4e-4d15-a277-1e8391f69e00/volumes" Nov 24 07:01:40 crc kubenswrapper[4799]: I1124 07:01:40.344280 4799 scope.go:117] "RemoveContainer" containerID="6050077eafe16fadc43066842743bfb05702c8fe964e70eb3f1c364f9fe56970" Nov 24 07:01:40 crc kubenswrapper[4799]: I1124 07:01:40.419159 4799 scope.go:117] "RemoveContainer" containerID="fb51fda68c12afdd26d507e56f5943c656540767b245f95ae678e206d8b591be" Nov 24 07:01:40 crc kubenswrapper[4799]: I1124 07:01:40.969740 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x5t8r"] Nov 24 07:01:40 crc kubenswrapper[4799]: I1124 07:01:40.970121 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x5t8r" podUID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" containerName="registry-server" containerID="cri-o://ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7" gracePeriod=2 Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.358302 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.446421 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-utilities\") pod \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.446588 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpq4z\" (UniqueName: \"kubernetes.io/projected/4e2791a7-9b30-427a-872e-cdb4967dc4b3-kube-api-access-mpq4z\") pod \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.446700 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-catalog-content\") pod \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\" (UID: \"4e2791a7-9b30-427a-872e-cdb4967dc4b3\") " Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.447481 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" event={"ID":"a45abe66-dc32-4b76-90d9-b821dba28bd5","Type":"ContainerStarted","Data":"4670794c6124d417c5d010d874352db9352d8040a5e00fadda96902d96a42379"} Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.447758 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.448268 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-utilities" (OuterVolumeSpecName: "utilities") pod "4e2791a7-9b30-427a-872e-cdb4967dc4b3" (UID: "4e2791a7-9b30-427a-872e-cdb4967dc4b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.449972 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z" event={"ID":"439904d9-d1e0-4733-8f08-db0f792fdcc2","Type":"ContainerStarted","Data":"3faffc422b24cfe628b0632825cb4b24081d22d973e7384e6fa2bcb494150b88"} Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.453231 4799 generic.go:334] "Generic (PLEG): container finished" podID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" containerID="ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7" exitCode=0 Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.453317 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5t8r" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.453338 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5t8r" event={"ID":"4e2791a7-9b30-427a-872e-cdb4967dc4b3","Type":"ContainerDied","Data":"ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7"} Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.453400 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5t8r" event={"ID":"4e2791a7-9b30-427a-872e-cdb4967dc4b3","Type":"ContainerDied","Data":"c28662a6b672c0c1f8b9578a3d5797532e6d43e381d1e0f6ea0e15c8a78f8799"} Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.453437 4799 scope.go:117] "RemoveContainer" containerID="ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.457093 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e2791a7-9b30-427a-872e-cdb4967dc4b3-kube-api-access-mpq4z" (OuterVolumeSpecName: "kube-api-access-mpq4z") pod "4e2791a7-9b30-427a-872e-cdb4967dc4b3" (UID: "4e2791a7-9b30-427a-872e-cdb4967dc4b3"). InnerVolumeSpecName "kube-api-access-mpq4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.476140 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" podStartSLOduration=2.136832246 podStartE2EDuration="11.476110186s" podCreationTimestamp="2025-11-24 07:01:30 +0000 UTC" firstStartedPulling="2025-11-24 07:01:31.107974685 +0000 UTC m=+836.763957159" lastFinishedPulling="2025-11-24 07:01:40.447252625 +0000 UTC m=+846.103235099" observedRunningTime="2025-11-24 07:01:41.473515463 +0000 UTC m=+847.129497957" watchObservedRunningTime="2025-11-24 07:01:41.476110186 +0000 UTC m=+847.132092670" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.496298 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-ktx6z" podStartSLOduration=3.700896996 podStartE2EDuration="8.496277927s" podCreationTimestamp="2025-11-24 07:01:33 +0000 UTC" firstStartedPulling="2025-11-24 07:01:35.648079476 +0000 UTC m=+841.304061970" lastFinishedPulling="2025-11-24 07:01:40.443460427 +0000 UTC m=+846.099442901" observedRunningTime="2025-11-24 07:01:41.491969805 +0000 UTC m=+847.147952299" watchObservedRunningTime="2025-11-24 07:01:41.496277927 +0000 UTC m=+847.152260411" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.500049 4799 scope.go:117] "RemoveContainer" containerID="1b3119925156ae14974e027ca914887526567e64f4b5e44e0730e5021283682c" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.502760 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e2791a7-9b30-427a-872e-cdb4967dc4b3" (UID: "4e2791a7-9b30-427a-872e-cdb4967dc4b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.521122 4799 scope.go:117] "RemoveContainer" containerID="0a0eafdeb538f21ab5c1a57460a2d358a1e67beef51090b7748071efc6e8c36d" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.548678 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.548722 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpq4z\" (UniqueName: \"kubernetes.io/projected/4e2791a7-9b30-427a-872e-cdb4967dc4b3-kube-api-access-mpq4z\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.548733 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2791a7-9b30-427a-872e-cdb4967dc4b3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.552365 4799 scope.go:117] "RemoveContainer" containerID="ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7" Nov 24 07:01:41 crc kubenswrapper[4799]: E1124 07:01:41.563619 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7\": container with ID starting with ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7 not found: ID does not exist" containerID="ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.563741 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7"} err="failed to get container status \"ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7\": rpc error: code = NotFound desc = could not find container \"ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7\": container with ID starting with ca85f65036101bcab5e6381496f56e127aaf1ac411e5f5b3a86fb49d2fb3dab7 not found: ID does not exist" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.563809 4799 scope.go:117] "RemoveContainer" containerID="1b3119925156ae14974e027ca914887526567e64f4b5e44e0730e5021283682c" Nov 24 07:01:41 crc kubenswrapper[4799]: E1124 07:01:41.565270 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b3119925156ae14974e027ca914887526567e64f4b5e44e0730e5021283682c\": container with ID starting with 1b3119925156ae14974e027ca914887526567e64f4b5e44e0730e5021283682c not found: ID does not exist" containerID="1b3119925156ae14974e027ca914887526567e64f4b5e44e0730e5021283682c" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.565359 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b3119925156ae14974e027ca914887526567e64f4b5e44e0730e5021283682c"} err="failed to get container status \"1b3119925156ae14974e027ca914887526567e64f4b5e44e0730e5021283682c\": rpc error: code = NotFound desc = could not find container \"1b3119925156ae14974e027ca914887526567e64f4b5e44e0730e5021283682c\": container with ID starting with 1b3119925156ae14974e027ca914887526567e64f4b5e44e0730e5021283682c not found: ID does not exist" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.565409 4799 scope.go:117] "RemoveContainer" containerID="0a0eafdeb538f21ab5c1a57460a2d358a1e67beef51090b7748071efc6e8c36d" Nov 24 07:01:41 crc kubenswrapper[4799]: E1124 07:01:41.567452 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a0eafdeb538f21ab5c1a57460a2d358a1e67beef51090b7748071efc6e8c36d\": container with ID starting with 0a0eafdeb538f21ab5c1a57460a2d358a1e67beef51090b7748071efc6e8c36d not found: ID does not exist" containerID="0a0eafdeb538f21ab5c1a57460a2d358a1e67beef51090b7748071efc6e8c36d" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.567491 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a0eafdeb538f21ab5c1a57460a2d358a1e67beef51090b7748071efc6e8c36d"} err="failed to get container status \"0a0eafdeb538f21ab5c1a57460a2d358a1e67beef51090b7748071efc6e8c36d\": rpc error: code = NotFound desc = could not find container \"0a0eafdeb538f21ab5c1a57460a2d358a1e67beef51090b7748071efc6e8c36d\": container with ID starting with 0a0eafdeb538f21ab5c1a57460a2d358a1e67beef51090b7748071efc6e8c36d not found: ID does not exist" Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.777270 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x5t8r"] Nov 24 07:01:41 crc kubenswrapper[4799]: I1124 07:01:41.788686 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x5t8r"] Nov 24 07:01:43 crc kubenswrapper[4799]: I1124 07:01:43.636897 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" path="/var/lib/kubelet/pods/4e2791a7-9b30-427a-872e-cdb4967dc4b3/volumes" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.288789 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-btwn9"] Nov 24 07:01:44 crc kubenswrapper[4799]: E1124 07:01:44.289236 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerName="extract-utilities" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289253 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerName="extract-utilities" Nov 24 07:01:44 crc kubenswrapper[4799]: E1124 07:01:44.289271 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" containerName="extract-utilities" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289278 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" containerName="extract-utilities" Nov 24 07:01:44 crc kubenswrapper[4799]: E1124 07:01:44.289288 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" containerName="registry-server" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289296 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" containerName="registry-server" Nov 24 07:01:44 crc kubenswrapper[4799]: E1124 07:01:44.289311 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" containerName="registry-server" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289366 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" containerName="registry-server" Nov 24 07:01:44 crc kubenswrapper[4799]: E1124 07:01:44.289378 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" containerName="extract-content" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289391 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" containerName="extract-content" Nov 24 07:01:44 crc kubenswrapper[4799]: E1124 07:01:44.289403 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" containerName="extract-utilities" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289409 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" containerName="extract-utilities" Nov 24 07:01:44 crc kubenswrapper[4799]: E1124 07:01:44.289423 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerName="registry-server" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289430 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerName="registry-server" Nov 24 07:01:44 crc kubenswrapper[4799]: E1124 07:01:44.289446 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" containerName="extract-content" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289454 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" containerName="extract-content" Nov 24 07:01:44 crc kubenswrapper[4799]: E1124 07:01:44.289463 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerName="extract-content" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289471 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerName="extract-content" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289607 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5935c614-4a8b-485f-84ff-a84cb8fe26df" containerName="registry-server" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289625 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e52f5c42-2d4e-4d15-a277-1e8391f69e00" containerName="registry-server" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.289641 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e2791a7-9b30-427a-872e-cdb4967dc4b3" containerName="registry-server" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.290273 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-btwn9" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.294673 4799 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-4lv99" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.304679 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-btwn9"] Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.393083 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf9nb\" (UniqueName: \"kubernetes.io/projected/71cb8206-339b-4aff-aa85-d009840a5fd6-kube-api-access-lf9nb\") pod \"cert-manager-86cb77c54b-btwn9\" (UID: \"71cb8206-339b-4aff-aa85-d009840a5fd6\") " pod="cert-manager/cert-manager-86cb77c54b-btwn9" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.393155 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71cb8206-339b-4aff-aa85-d009840a5fd6-bound-sa-token\") pod \"cert-manager-86cb77c54b-btwn9\" (UID: \"71cb8206-339b-4aff-aa85-d009840a5fd6\") " pod="cert-manager/cert-manager-86cb77c54b-btwn9" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.493599 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71cb8206-339b-4aff-aa85-d009840a5fd6-bound-sa-token\") pod \"cert-manager-86cb77c54b-btwn9\" (UID: \"71cb8206-339b-4aff-aa85-d009840a5fd6\") " pod="cert-manager/cert-manager-86cb77c54b-btwn9" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.493685 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf9nb\" (UniqueName: \"kubernetes.io/projected/71cb8206-339b-4aff-aa85-d009840a5fd6-kube-api-access-lf9nb\") pod \"cert-manager-86cb77c54b-btwn9\" (UID: \"71cb8206-339b-4aff-aa85-d009840a5fd6\") " pod="cert-manager/cert-manager-86cb77c54b-btwn9" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.516337 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71cb8206-339b-4aff-aa85-d009840a5fd6-bound-sa-token\") pod \"cert-manager-86cb77c54b-btwn9\" (UID: \"71cb8206-339b-4aff-aa85-d009840a5fd6\") " pod="cert-manager/cert-manager-86cb77c54b-btwn9" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.517673 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf9nb\" (UniqueName: \"kubernetes.io/projected/71cb8206-339b-4aff-aa85-d009840a5fd6-kube-api-access-lf9nb\") pod \"cert-manager-86cb77c54b-btwn9\" (UID: \"71cb8206-339b-4aff-aa85-d009840a5fd6\") " pod="cert-manager/cert-manager-86cb77c54b-btwn9" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.622440 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-btwn9" Nov 24 07:01:44 crc kubenswrapper[4799]: I1124 07:01:44.873988 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-btwn9"] Nov 24 07:01:45 crc kubenswrapper[4799]: I1124 07:01:45.505121 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-btwn9" event={"ID":"71cb8206-339b-4aff-aa85-d009840a5fd6","Type":"ContainerStarted","Data":"668060184cea3d33e30c960744e99f6f1d9afc6241d433bfe94a0b16acaae8cd"} Nov 24 07:01:45 crc kubenswrapper[4799]: I1124 07:01:45.505562 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-btwn9" event={"ID":"71cb8206-339b-4aff-aa85-d009840a5fd6","Type":"ContainerStarted","Data":"8f5770a69acf688478cdc3121934ff9f152c5a26b06e2f3fce64bc9ed570149d"} Nov 24 07:01:45 crc kubenswrapper[4799]: I1124 07:01:45.537785 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-btwn9" podStartSLOduration=1.537744553 podStartE2EDuration="1.537744553s" podCreationTimestamp="2025-11-24 07:01:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:01:45.531172317 +0000 UTC m=+851.187154831" watchObservedRunningTime="2025-11-24 07:01:45.537744553 +0000 UTC m=+851.193727067" Nov 24 07:01:45 crc kubenswrapper[4799]: I1124 07:01:45.614256 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-58pnr" Nov 24 07:01:47 crc kubenswrapper[4799]: E1124 07:01:47.376745 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/systemd-tmpfiles-clean.service\": RecentStats: unable to find data in memory cache]" Nov 24 07:01:48 crc kubenswrapper[4799]: I1124 07:01:48.785151 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-wwx8z"] Nov 24 07:01:48 crc kubenswrapper[4799]: I1124 07:01:48.786776 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wwx8z" Nov 24 07:01:48 crc kubenswrapper[4799]: I1124 07:01:48.790312 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-bt8wr" Nov 24 07:01:48 crc kubenswrapper[4799]: I1124 07:01:48.791512 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 07:01:48 crc kubenswrapper[4799]: I1124 07:01:48.794142 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 07:01:48 crc kubenswrapper[4799]: I1124 07:01:48.804255 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wwx8z"] Nov 24 07:01:48 crc kubenswrapper[4799]: I1124 07:01:48.963200 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc5pz\" (UniqueName: \"kubernetes.io/projected/984498a3-832d-43bc-ae72-4c18fbdab069-kube-api-access-gc5pz\") pod \"openstack-operator-index-wwx8z\" (UID: \"984498a3-832d-43bc-ae72-4c18fbdab069\") " pod="openstack-operators/openstack-operator-index-wwx8z" Nov 24 07:01:49 crc kubenswrapper[4799]: I1124 07:01:49.064355 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc5pz\" (UniqueName: \"kubernetes.io/projected/984498a3-832d-43bc-ae72-4c18fbdab069-kube-api-access-gc5pz\") pod \"openstack-operator-index-wwx8z\" (UID: \"984498a3-832d-43bc-ae72-4c18fbdab069\") " pod="openstack-operators/openstack-operator-index-wwx8z" Nov 24 07:01:49 crc kubenswrapper[4799]: I1124 07:01:49.098728 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc5pz\" (UniqueName: \"kubernetes.io/projected/984498a3-832d-43bc-ae72-4c18fbdab069-kube-api-access-gc5pz\") pod \"openstack-operator-index-wwx8z\" (UID: \"984498a3-832d-43bc-ae72-4c18fbdab069\") " pod="openstack-operators/openstack-operator-index-wwx8z" Nov 24 07:01:49 crc kubenswrapper[4799]: I1124 07:01:49.114136 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wwx8z" Nov 24 07:01:49 crc kubenswrapper[4799]: I1124 07:01:49.572797 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wwx8z"] Nov 24 07:01:49 crc kubenswrapper[4799]: W1124 07:01:49.584021 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod984498a3_832d_43bc_ae72_4c18fbdab069.slice/crio-d831bff4522c66cfc70f704b1209f0deb6afe04aa64674110eb75438a74423e9 WatchSource:0}: Error finding container d831bff4522c66cfc70f704b1209f0deb6afe04aa64674110eb75438a74423e9: Status 404 returned error can't find the container with id d831bff4522c66cfc70f704b1209f0deb6afe04aa64674110eb75438a74423e9 Nov 24 07:01:50 crc kubenswrapper[4799]: I1124 07:01:50.564214 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wwx8z" event={"ID":"984498a3-832d-43bc-ae72-4c18fbdab069","Type":"ContainerStarted","Data":"5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042"} Nov 24 07:01:50 crc kubenswrapper[4799]: I1124 07:01:50.565166 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wwx8z" event={"ID":"984498a3-832d-43bc-ae72-4c18fbdab069","Type":"ContainerStarted","Data":"d831bff4522c66cfc70f704b1209f0deb6afe04aa64674110eb75438a74423e9"} Nov 24 07:01:52 crc kubenswrapper[4799]: I1124 07:01:52.165276 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-wwx8z" podStartSLOduration=3.417130375 podStartE2EDuration="4.165232746s" podCreationTimestamp="2025-11-24 07:01:48 +0000 UTC" firstStartedPulling="2025-11-24 07:01:49.592593707 +0000 UTC m=+855.248576231" lastFinishedPulling="2025-11-24 07:01:50.340696118 +0000 UTC m=+855.996678602" observedRunningTime="2025-11-24 07:01:50.592403958 +0000 UTC m=+856.248386432" watchObservedRunningTime="2025-11-24 07:01:52.165232746 +0000 UTC m=+857.821215260" Nov 24 07:01:52 crc kubenswrapper[4799]: I1124 07:01:52.166668 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-wwx8z"] Nov 24 07:01:52 crc kubenswrapper[4799]: I1124 07:01:52.582892 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-wwx8z" podUID="984498a3-832d-43bc-ae72-4c18fbdab069" containerName="registry-server" containerID="cri-o://5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042" gracePeriod=2 Nov 24 07:01:52 crc kubenswrapper[4799]: I1124 07:01:52.779757 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-4k5qm"] Nov 24 07:01:52 crc kubenswrapper[4799]: I1124 07:01:52.781916 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4k5qm" Nov 24 07:01:52 crc kubenswrapper[4799]: I1124 07:01:52.792384 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4k5qm"] Nov 24 07:01:52 crc kubenswrapper[4799]: I1124 07:01:52.827151 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt648\" (UniqueName: \"kubernetes.io/projected/3d678199-c527-4db6-8e3a-39ba2e59c03a-kube-api-access-lt648\") pod \"openstack-operator-index-4k5qm\" (UID: \"3d678199-c527-4db6-8e3a-39ba2e59c03a\") " pod="openstack-operators/openstack-operator-index-4k5qm" Nov 24 07:01:52 crc kubenswrapper[4799]: I1124 07:01:52.928165 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt648\" (UniqueName: \"kubernetes.io/projected/3d678199-c527-4db6-8e3a-39ba2e59c03a-kube-api-access-lt648\") pod \"openstack-operator-index-4k5qm\" (UID: \"3d678199-c527-4db6-8e3a-39ba2e59c03a\") " pod="openstack-operators/openstack-operator-index-4k5qm" Nov 24 07:01:52 crc kubenswrapper[4799]: I1124 07:01:52.952625 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt648\" (UniqueName: \"kubernetes.io/projected/3d678199-c527-4db6-8e3a-39ba2e59c03a-kube-api-access-lt648\") pod \"openstack-operator-index-4k5qm\" (UID: \"3d678199-c527-4db6-8e3a-39ba2e59c03a\") " pod="openstack-operators/openstack-operator-index-4k5qm" Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.027050 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wwx8z" Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.103800 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4k5qm" Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.130207 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc5pz\" (UniqueName: \"kubernetes.io/projected/984498a3-832d-43bc-ae72-4c18fbdab069-kube-api-access-gc5pz\") pod \"984498a3-832d-43bc-ae72-4c18fbdab069\" (UID: \"984498a3-832d-43bc-ae72-4c18fbdab069\") " Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.134331 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/984498a3-832d-43bc-ae72-4c18fbdab069-kube-api-access-gc5pz" (OuterVolumeSpecName: "kube-api-access-gc5pz") pod "984498a3-832d-43bc-ae72-4c18fbdab069" (UID: "984498a3-832d-43bc-ae72-4c18fbdab069"). InnerVolumeSpecName "kube-api-access-gc5pz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.232840 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc5pz\" (UniqueName: \"kubernetes.io/projected/984498a3-832d-43bc-ae72-4c18fbdab069-kube-api-access-gc5pz\") on node \"crc\" DevicePath \"\"" Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.566349 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4k5qm"] Nov 24 07:01:53 crc kubenswrapper[4799]: W1124 07:01:53.567579 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d678199_c527_4db6_8e3a_39ba2e59c03a.slice/crio-5ddd0981c887dea321cdaa4546430fc083881ef0eade4c35922e1067605c2d37 WatchSource:0}: Error finding container 5ddd0981c887dea321cdaa4546430fc083881ef0eade4c35922e1067605c2d37: Status 404 returned error can't find the container with id 5ddd0981c887dea321cdaa4546430fc083881ef0eade4c35922e1067605c2d37 Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.593349 4799 generic.go:334] "Generic (PLEG): container finished" podID="984498a3-832d-43bc-ae72-4c18fbdab069" containerID="5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042" exitCode=0 Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.593440 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wwx8z" event={"ID":"984498a3-832d-43bc-ae72-4c18fbdab069","Type":"ContainerDied","Data":"5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042"} Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.593474 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wwx8z" event={"ID":"984498a3-832d-43bc-ae72-4c18fbdab069","Type":"ContainerDied","Data":"d831bff4522c66cfc70f704b1209f0deb6afe04aa64674110eb75438a74423e9"} Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.593493 4799 scope.go:117] "RemoveContainer" containerID="5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042" Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.593590 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wwx8z" Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.597051 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4k5qm" event={"ID":"3d678199-c527-4db6-8e3a-39ba2e59c03a","Type":"ContainerStarted","Data":"5ddd0981c887dea321cdaa4546430fc083881ef0eade4c35922e1067605c2d37"} Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.617785 4799 scope.go:117] "RemoveContainer" containerID="5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042" Nov 24 07:01:53 crc kubenswrapper[4799]: E1124 07:01:53.618665 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042\": container with ID starting with 5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042 not found: ID does not exist" containerID="5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042" Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.618738 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042"} err="failed to get container status \"5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042\": rpc error: code = NotFound desc = could not find container \"5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042\": container with ID starting with 5842ffb5f62b70c1c406f4ff6fc296cedade3c2ba78ca8b0e2b8380fe7989042 not found: ID does not exist" Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.628674 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-wwx8z"] Nov 24 07:01:53 crc kubenswrapper[4799]: I1124 07:01:53.644304 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-wwx8z"] Nov 24 07:01:54 crc kubenswrapper[4799]: I1124 07:01:54.611442 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4k5qm" event={"ID":"3d678199-c527-4db6-8e3a-39ba2e59c03a","Type":"ContainerStarted","Data":"cca1e40333ca1ca0adaf70915fec222a9b8b1db41990ac33befac8e30e2e37f9"} Nov 24 07:01:54 crc kubenswrapper[4799]: I1124 07:01:54.642196 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-4k5qm" podStartSLOduration=2.208311936 podStartE2EDuration="2.642165638s" podCreationTimestamp="2025-11-24 07:01:52 +0000 UTC" firstStartedPulling="2025-11-24 07:01:53.573240713 +0000 UTC m=+859.229223187" lastFinishedPulling="2025-11-24 07:01:54.007094415 +0000 UTC m=+859.663076889" observedRunningTime="2025-11-24 07:01:54.635038467 +0000 UTC m=+860.291021001" watchObservedRunningTime="2025-11-24 07:01:54.642165638 +0000 UTC m=+860.298148122" Nov 24 07:01:55 crc kubenswrapper[4799]: I1124 07:01:55.642568 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="984498a3-832d-43bc-ae72-4c18fbdab069" path="/var/lib/kubelet/pods/984498a3-832d-43bc-ae72-4c18fbdab069/volumes" Nov 24 07:02:03 crc kubenswrapper[4799]: I1124 07:02:03.104178 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-4k5qm" Nov 24 07:02:03 crc kubenswrapper[4799]: I1124 07:02:03.105141 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-4k5qm" Nov 24 07:02:03 crc kubenswrapper[4799]: I1124 07:02:03.143480 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-4k5qm" Nov 24 07:02:03 crc kubenswrapper[4799]: I1124 07:02:03.736317 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-4k5qm" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.425820 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs"] Nov 24 07:02:11 crc kubenswrapper[4799]: E1124 07:02:11.427074 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984498a3-832d-43bc-ae72-4c18fbdab069" containerName="registry-server" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.427097 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="984498a3-832d-43bc-ae72-4c18fbdab069" containerName="registry-server" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.427298 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="984498a3-832d-43bc-ae72-4c18fbdab069" containerName="registry-server" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.428725 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.431819 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8sm2d" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.440818 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs"] Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.580477 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.580659 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcnq2\" (UniqueName: \"kubernetes.io/projected/3a457c47-44d1-4246-a71b-3c1dc628564e-kube-api-access-zcnq2\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.580789 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.682980 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcnq2\" (UniqueName: \"kubernetes.io/projected/3a457c47-44d1-4246-a71b-3c1dc628564e-kube-api-access-zcnq2\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.683119 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.683274 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.685011 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.685080 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.709065 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcnq2\" (UniqueName: \"kubernetes.io/projected/3a457c47-44d1-4246-a71b-3c1dc628564e-kube-api-access-zcnq2\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:11 crc kubenswrapper[4799]: I1124 07:02:11.750255 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:12 crc kubenswrapper[4799]: I1124 07:02:12.257573 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs"] Nov 24 07:02:12 crc kubenswrapper[4799]: I1124 07:02:12.763137 4799 generic.go:334] "Generic (PLEG): container finished" podID="3a457c47-44d1-4246-a71b-3c1dc628564e" containerID="c8660677e412fb891eb8bac754050f74203d43e5a3cc7a49e08fa8aaa30d6bc4" exitCode=0 Nov 24 07:02:12 crc kubenswrapper[4799]: I1124 07:02:12.763266 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" event={"ID":"3a457c47-44d1-4246-a71b-3c1dc628564e","Type":"ContainerDied","Data":"c8660677e412fb891eb8bac754050f74203d43e5a3cc7a49e08fa8aaa30d6bc4"} Nov 24 07:02:12 crc kubenswrapper[4799]: I1124 07:02:12.763675 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" event={"ID":"3a457c47-44d1-4246-a71b-3c1dc628564e","Type":"ContainerStarted","Data":"794997998fd64b74d2c81245356ce5bface9e2fdacb3b8ff00e59c144f6b526a"} Nov 24 07:02:13 crc kubenswrapper[4799]: I1124 07:02:13.772789 4799 generic.go:334] "Generic (PLEG): container finished" podID="3a457c47-44d1-4246-a71b-3c1dc628564e" containerID="7320d8f8ba42ca596d07ef2115be5846ca94db3cb87950ed98c7c626f20a50c6" exitCode=0 Nov 24 07:02:13 crc kubenswrapper[4799]: I1124 07:02:13.772897 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" event={"ID":"3a457c47-44d1-4246-a71b-3c1dc628564e","Type":"ContainerDied","Data":"7320d8f8ba42ca596d07ef2115be5846ca94db3cb87950ed98c7c626f20a50c6"} Nov 24 07:02:14 crc kubenswrapper[4799]: I1124 07:02:14.825906 4799 generic.go:334] "Generic (PLEG): container finished" podID="3a457c47-44d1-4246-a71b-3c1dc628564e" containerID="0a95d2d8d51a8fc784c2886a1e257738e1cfc897b526be1e05938b79c440c319" exitCode=0 Nov 24 07:02:14 crc kubenswrapper[4799]: I1124 07:02:14.825980 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" event={"ID":"3a457c47-44d1-4246-a71b-3c1dc628564e","Type":"ContainerDied","Data":"0a95d2d8d51a8fc784c2886a1e257738e1cfc897b526be1e05938b79c440c319"} Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.225024 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.370212 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-bundle\") pod \"3a457c47-44d1-4246-a71b-3c1dc628564e\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.370295 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-util\") pod \"3a457c47-44d1-4246-a71b-3c1dc628564e\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.370482 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcnq2\" (UniqueName: \"kubernetes.io/projected/3a457c47-44d1-4246-a71b-3c1dc628564e-kube-api-access-zcnq2\") pod \"3a457c47-44d1-4246-a71b-3c1dc628564e\" (UID: \"3a457c47-44d1-4246-a71b-3c1dc628564e\") " Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.372350 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-bundle" (OuterVolumeSpecName: "bundle") pod "3a457c47-44d1-4246-a71b-3c1dc628564e" (UID: "3a457c47-44d1-4246-a71b-3c1dc628564e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.376218 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a457c47-44d1-4246-a71b-3c1dc628564e-kube-api-access-zcnq2" (OuterVolumeSpecName: "kube-api-access-zcnq2") pod "3a457c47-44d1-4246-a71b-3c1dc628564e" (UID: "3a457c47-44d1-4246-a71b-3c1dc628564e"). InnerVolumeSpecName "kube-api-access-zcnq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.386212 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-util" (OuterVolumeSpecName: "util") pod "3a457c47-44d1-4246-a71b-3c1dc628564e" (UID: "3a457c47-44d1-4246-a71b-3c1dc628564e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.472936 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcnq2\" (UniqueName: \"kubernetes.io/projected/3a457c47-44d1-4246-a71b-3c1dc628564e-kube-api-access-zcnq2\") on node \"crc\" DevicePath \"\"" Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.472993 4799 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.473012 4799 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a457c47-44d1-4246-a71b-3c1dc628564e-util\") on node \"crc\" DevicePath \"\"" Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.863947 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" event={"ID":"3a457c47-44d1-4246-a71b-3c1dc628564e","Type":"ContainerDied","Data":"794997998fd64b74d2c81245356ce5bface9e2fdacb3b8ff00e59c144f6b526a"} Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.864011 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="794997998fd64b74d2c81245356ce5bface9e2fdacb3b8ff00e59c144f6b526a" Nov 24 07:02:16 crc kubenswrapper[4799]: I1124 07:02:16.864050 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs" Nov 24 07:02:20 crc kubenswrapper[4799]: I1124 07:02:20.401256 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:02:20 crc kubenswrapper[4799]: I1124 07:02:20.401901 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.192579 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm"] Nov 24 07:02:24 crc kubenswrapper[4799]: E1124 07:02:24.193512 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a457c47-44d1-4246-a71b-3c1dc628564e" containerName="extract" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.193548 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a457c47-44d1-4246-a71b-3c1dc628564e" containerName="extract" Nov 24 07:02:24 crc kubenswrapper[4799]: E1124 07:02:24.193563 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a457c47-44d1-4246-a71b-3c1dc628564e" containerName="util" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.193570 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a457c47-44d1-4246-a71b-3c1dc628564e" containerName="util" Nov 24 07:02:24 crc kubenswrapper[4799]: E1124 07:02:24.193578 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a457c47-44d1-4246-a71b-3c1dc628564e" containerName="pull" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.193586 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a457c47-44d1-4246-a71b-3c1dc628564e" containerName="pull" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.193730 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a457c47-44d1-4246-a71b-3c1dc628564e" containerName="extract" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.195997 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.202246 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-f9mfp" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.243930 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm"] Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.314974 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grlxh\" (UniqueName: \"kubernetes.io/projected/98ea352f-7485-4cdb-9ec9-bb208357fefe-kube-api-access-grlxh\") pod \"openstack-operator-controller-operator-8486c7f98b-g6rfm\" (UID: \"98ea352f-7485-4cdb-9ec9-bb208357fefe\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.416803 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grlxh\" (UniqueName: \"kubernetes.io/projected/98ea352f-7485-4cdb-9ec9-bb208357fefe-kube-api-access-grlxh\") pod \"openstack-operator-controller-operator-8486c7f98b-g6rfm\" (UID: \"98ea352f-7485-4cdb-9ec9-bb208357fefe\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.451155 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grlxh\" (UniqueName: \"kubernetes.io/projected/98ea352f-7485-4cdb-9ec9-bb208357fefe-kube-api-access-grlxh\") pod \"openstack-operator-controller-operator-8486c7f98b-g6rfm\" (UID: \"98ea352f-7485-4cdb-9ec9-bb208357fefe\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.516269 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm" Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.776546 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm"] Nov 24 07:02:24 crc kubenswrapper[4799]: I1124 07:02:24.930818 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm" event={"ID":"98ea352f-7485-4cdb-9ec9-bb208357fefe","Type":"ContainerStarted","Data":"62cdaa21678d3c1e633c58df01b3496584645ebae5128b56ed0201b872739ced"} Nov 24 07:02:29 crc kubenswrapper[4799]: I1124 07:02:29.975683 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm" event={"ID":"98ea352f-7485-4cdb-9ec9-bb208357fefe","Type":"ContainerStarted","Data":"b9f03cb5c3948514b100febe524ab3fa58058adaa06f7c0e547164a91c401b70"} Nov 24 07:02:31 crc kubenswrapper[4799]: I1124 07:02:31.993624 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm" event={"ID":"98ea352f-7485-4cdb-9ec9-bb208357fefe","Type":"ContainerStarted","Data":"8217df90fdfcb6d5f6a02f89919a35775e3289c56b2a3d79c6b9ea6f7fe89f12"} Nov 24 07:02:31 crc kubenswrapper[4799]: I1124 07:02:31.994164 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm" Nov 24 07:02:32 crc kubenswrapper[4799]: I1124 07:02:32.036653 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm" podStartSLOduration=1.383187018 podStartE2EDuration="8.036622996s" podCreationTimestamp="2025-11-24 07:02:24 +0000 UTC" firstStartedPulling="2025-11-24 07:02:24.794164737 +0000 UTC m=+890.450147211" lastFinishedPulling="2025-11-24 07:02:31.447600715 +0000 UTC m=+897.103583189" observedRunningTime="2025-11-24 07:02:32.026715815 +0000 UTC m=+897.682698319" watchObservedRunningTime="2025-11-24 07:02:32.036622996 +0000 UTC m=+897.692605510" Nov 24 07:02:34 crc kubenswrapper[4799]: I1124 07:02:34.708633 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-g6rfm" Nov 24 07:02:50 crc kubenswrapper[4799]: I1124 07:02:50.401467 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:02:50 crc kubenswrapper[4799]: I1124 07:02:50.403910 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.185377 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.187656 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.194826 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hphvt" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.204156 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.220108 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.221432 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.224936 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-87478" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.231263 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.279829 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.281032 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.284534 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-zvpdt" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.298603 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.299717 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.315362 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-vj8q2" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.321115 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khqln\" (UniqueName: \"kubernetes.io/projected/f1284cfd-b954-4b2d-8e3c-2e57bb52ca27-kube-api-access-khqln\") pod \"barbican-operator-controller-manager-7768f8c84f-kzmqt\" (UID: \"f1284cfd-b954-4b2d-8e3c-2e57bb52ca27\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.321168 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqrpd\" (UniqueName: \"kubernetes.io/projected/52251a10-6203-43da-af96-94f78f0edacc-kube-api-access-sqrpd\") pod \"cinder-operator-controller-manager-6d8fd67bf7-w7jld\" (UID: \"52251a10-6203-43da-af96-94f78f0edacc\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.323453 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.327021 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.340507 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.342171 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.345660 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-cl9dw" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.371542 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.372525 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.373265 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.373724 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.382275 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.382566 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-48xxz" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.382735 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-rbj2n" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.391485 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.401557 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.409259 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.432075 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khqln\" (UniqueName: \"kubernetes.io/projected/f1284cfd-b954-4b2d-8e3c-2e57bb52ca27-kube-api-access-khqln\") pod \"barbican-operator-controller-manager-7768f8c84f-kzmqt\" (UID: \"f1284cfd-b954-4b2d-8e3c-2e57bb52ca27\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.432168 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mjq6\" (UniqueName: \"kubernetes.io/projected/4866479c-d0b9-4366-8ab4-f041573d0d35-kube-api-access-5mjq6\") pod \"glance-operator-controller-manager-8667fbf6f6-9prrl\" (UID: \"4866479c-d0b9-4366-8ab4-f041573d0d35\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.432205 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqrpd\" (UniqueName: \"kubernetes.io/projected/52251a10-6203-43da-af96-94f78f0edacc-kube-api-access-sqrpd\") pod \"cinder-operator-controller-manager-6d8fd67bf7-w7jld\" (UID: \"52251a10-6203-43da-af96-94f78f0edacc\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.432292 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xfhf\" (UniqueName: \"kubernetes.io/projected/3fec8392-f37a-4f4a-bdec-696bf7b02cd1-kube-api-access-4xfhf\") pod \"designate-operator-controller-manager-56dfb6b67f-rxlcx\" (UID: \"3fec8392-f37a-4f4a-bdec-696bf7b02cd1\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.464827 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.535503 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhkcl\" (UniqueName: \"kubernetes.io/projected/9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c-kube-api-access-xhkcl\") pod \"infra-operator-controller-manager-769d9c7585-s7z8n\" (UID: \"9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.536009 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mjq6\" (UniqueName: \"kubernetes.io/projected/4866479c-d0b9-4366-8ab4-f041573d0d35-kube-api-access-5mjq6\") pod \"glance-operator-controller-manager-8667fbf6f6-9prrl\" (UID: \"4866479c-d0b9-4366-8ab4-f041573d0d35\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.536049 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk9kj\" (UniqueName: \"kubernetes.io/projected/9f756dcc-daa5-4ed3-a4af-3bbf3c9f186c-kube-api-access-bk9kj\") pod \"horizon-operator-controller-manager-5d86b44686-9x9d2\" (UID: \"9f756dcc-daa5-4ed3-a4af-3bbf3c9f186c\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.536142 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv4vl\" (UniqueName: \"kubernetes.io/projected/af65b87f-5fe7-4ade-873b-e4d7040df219-kube-api-access-xv4vl\") pod \"heat-operator-controller-manager-bf4c6585d-2zmvd\" (UID: \"af65b87f-5fe7-4ade-873b-e4d7040df219\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.536161 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c-cert\") pod \"infra-operator-controller-manager-769d9c7585-s7z8n\" (UID: \"9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.536193 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xfhf\" (UniqueName: \"kubernetes.io/projected/3fec8392-f37a-4f4a-bdec-696bf7b02cd1-kube-api-access-4xfhf\") pod \"designate-operator-controller-manager-56dfb6b67f-rxlcx\" (UID: \"3fec8392-f37a-4f4a-bdec-696bf7b02cd1\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.537146 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.539796 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqrpd\" (UniqueName: \"kubernetes.io/projected/52251a10-6203-43da-af96-94f78f0edacc-kube-api-access-sqrpd\") pod \"cinder-operator-controller-manager-6d8fd67bf7-w7jld\" (UID: \"52251a10-6203-43da-af96-94f78f0edacc\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.540317 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-4z7t8" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.559718 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.560356 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.561519 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.568696 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khqln\" (UniqueName: \"kubernetes.io/projected/f1284cfd-b954-4b2d-8e3c-2e57bb52ca27-kube-api-access-khqln\") pod \"barbican-operator-controller-manager-7768f8c84f-kzmqt\" (UID: \"f1284cfd-b954-4b2d-8e3c-2e57bb52ca27\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.569452 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-mgkvl" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.571895 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xfhf\" (UniqueName: \"kubernetes.io/projected/3fec8392-f37a-4f4a-bdec-696bf7b02cd1-kube-api-access-4xfhf\") pod \"designate-operator-controller-manager-56dfb6b67f-rxlcx\" (UID: \"3fec8392-f37a-4f4a-bdec-696bf7b02cd1\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.572534 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mjq6\" (UniqueName: \"kubernetes.io/projected/4866479c-d0b9-4366-8ab4-f041573d0d35-kube-api-access-5mjq6\") pod \"glance-operator-controller-manager-8667fbf6f6-9prrl\" (UID: \"4866479c-d0b9-4366-8ab4-f041573d0d35\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.582934 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.604930 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.605548 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.620942 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.653598 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhkcl\" (UniqueName: \"kubernetes.io/projected/9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c-kube-api-access-xhkcl\") pod \"infra-operator-controller-manager-769d9c7585-s7z8n\" (UID: \"9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.653666 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk9kj\" (UniqueName: \"kubernetes.io/projected/9f756dcc-daa5-4ed3-a4af-3bbf3c9f186c-kube-api-access-bk9kj\") pod \"horizon-operator-controller-manager-5d86b44686-9x9d2\" (UID: \"9f756dcc-daa5-4ed3-a4af-3bbf3c9f186c\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.653696 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9nnr\" (UniqueName: \"kubernetes.io/projected/d6a55aa5-3e1c-49e9-bd56-48f8497c1b49-kube-api-access-b9nnr\") pod \"ironic-operator-controller-manager-5c75d7c94b-mjdkn\" (UID: \"d6a55aa5-3e1c-49e9-bd56-48f8497c1b49\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.653727 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv4vl\" (UniqueName: \"kubernetes.io/projected/af65b87f-5fe7-4ade-873b-e4d7040df219-kube-api-access-xv4vl\") pod \"heat-operator-controller-manager-bf4c6585d-2zmvd\" (UID: \"af65b87f-5fe7-4ade-873b-e4d7040df219\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.653747 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c-cert\") pod \"infra-operator-controller-manager-769d9c7585-s7z8n\" (UID: \"9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.653769 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7gx5\" (UniqueName: \"kubernetes.io/projected/32b50dc4-f35a-419b-9a18-9ff05f4c826c-kube-api-access-r7gx5\") pod \"keystone-operator-controller-manager-7879fb76fd-nwftw\" (UID: \"32b50dc4-f35a-419b-9a18-9ff05f4c826c\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.664295 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.665907 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.668648 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c-cert\") pod \"infra-operator-controller-manager-769d9c7585-s7z8n\" (UID: \"9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.680731 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-dqhwx" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.701695 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.703121 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.703656 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhkcl\" (UniqueName: \"kubernetes.io/projected/9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c-kube-api-access-xhkcl\") pod \"infra-operator-controller-manager-769d9c7585-s7z8n\" (UID: \"9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.704268 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk9kj\" (UniqueName: \"kubernetes.io/projected/9f756dcc-daa5-4ed3-a4af-3bbf3c9f186c-kube-api-access-bk9kj\") pod \"horizon-operator-controller-manager-5d86b44686-9x9d2\" (UID: \"9f756dcc-daa5-4ed3-a4af-3bbf3c9f186c\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.704631 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv4vl\" (UniqueName: \"kubernetes.io/projected/af65b87f-5fe7-4ade-873b-e4d7040df219-kube-api-access-xv4vl\") pod \"heat-operator-controller-manager-bf4c6585d-2zmvd\" (UID: \"af65b87f-5fe7-4ade-873b-e4d7040df219\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.711611 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.717405 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-knmw2" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.726939 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.728631 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.733455 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-w8tt6" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.736250 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.754879 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9nnr\" (UniqueName: \"kubernetes.io/projected/d6a55aa5-3e1c-49e9-bd56-48f8497c1b49-kube-api-access-b9nnr\") pod \"ironic-operator-controller-manager-5c75d7c94b-mjdkn\" (UID: \"d6a55aa5-3e1c-49e9-bd56-48f8497c1b49\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.754961 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7gx5\" (UniqueName: \"kubernetes.io/projected/32b50dc4-f35a-419b-9a18-9ff05f4c826c-kube-api-access-r7gx5\") pod \"keystone-operator-controller-manager-7879fb76fd-nwftw\" (UID: \"32b50dc4-f35a-419b-9a18-9ff05f4c826c\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.758278 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.779425 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.780752 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.795782 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-tfpbn" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.801158 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.808561 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7gx5\" (UniqueName: \"kubernetes.io/projected/32b50dc4-f35a-419b-9a18-9ff05f4c826c-kube-api-access-r7gx5\") pod \"keystone-operator-controller-manager-7879fb76fd-nwftw\" (UID: \"32b50dc4-f35a-419b-9a18-9ff05f4c826c\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.820358 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.825019 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9nnr\" (UniqueName: \"kubernetes.io/projected/d6a55aa5-3e1c-49e9-bd56-48f8497c1b49-kube-api-access-b9nnr\") pod \"ironic-operator-controller-manager-5c75d7c94b-mjdkn\" (UID: \"d6a55aa5-3e1c-49e9-bd56-48f8497c1b49\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.846891 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.860292 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghrn7\" (UniqueName: \"kubernetes.io/projected/3b5de07d-ee5e-492a-9b9d-26de9837ca80-kube-api-access-ghrn7\") pod \"nova-operator-controller-manager-86d796d84d-5v5lr\" (UID: \"3b5de07d-ee5e-492a-9b9d-26de9837ca80\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.860358 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2jgc\" (UniqueName: \"kubernetes.io/projected/6e6f9937-437a-4f97-a8ea-e486ec9a448c-kube-api-access-b2jgc\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-kxlc9\" (UID: \"6e6f9937-437a-4f97-a8ea-e486ec9a448c\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.860402 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqwjp\" (UniqueName: \"kubernetes.io/projected/95afe811-a25f-4b89-9928-ee261693091b-kube-api-access-rqwjp\") pod \"manila-operator-controller-manager-7bb88cb858-2lccf\" (UID: \"95afe811-a25f-4b89-9928-ee261693091b\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.860452 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm2zl\" (UniqueName: \"kubernetes.io/projected/8a1443d1-4bea-481d-ad92-c92e7f5ac7cd-kube-api-access-tm2zl\") pod \"neutron-operator-controller-manager-66b7d6f598-xgq2g\" (UID: \"8a1443d1-4bea-481d-ad92-c92e7f5ac7cd\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.873568 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.903179 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.905683 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.925178 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-rql9x" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.935213 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.936748 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.945694 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-8qrdd" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.956687 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc"] Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.963212 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.967803 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghrn7\" (UniqueName: \"kubernetes.io/projected/3b5de07d-ee5e-492a-9b9d-26de9837ca80-kube-api-access-ghrn7\") pod \"nova-operator-controller-manager-86d796d84d-5v5lr\" (UID: \"3b5de07d-ee5e-492a-9b9d-26de9837ca80\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.967952 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2jgc\" (UniqueName: \"kubernetes.io/projected/6e6f9937-437a-4f97-a8ea-e486ec9a448c-kube-api-access-b2jgc\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-kxlc9\" (UID: \"6e6f9937-437a-4f97-a8ea-e486ec9a448c\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.968004 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqwjp\" (UniqueName: \"kubernetes.io/projected/95afe811-a25f-4b89-9928-ee261693091b-kube-api-access-rqwjp\") pod \"manila-operator-controller-manager-7bb88cb858-2lccf\" (UID: \"95afe811-a25f-4b89-9928-ee261693091b\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.968128 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm2zl\" (UniqueName: \"kubernetes.io/projected/8a1443d1-4bea-481d-ad92-c92e7f5ac7cd-kube-api-access-tm2zl\") pod \"neutron-operator-controller-manager-66b7d6f598-xgq2g\" (UID: \"8a1443d1-4bea-481d-ad92-c92e7f5ac7cd\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.994096 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" Nov 24 07:03:08 crc kubenswrapper[4799]: I1124 07:03:08.994747 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.006526 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqwjp\" (UniqueName: \"kubernetes.io/projected/95afe811-a25f-4b89-9928-ee261693091b-kube-api-access-rqwjp\") pod \"manila-operator-controller-manager-7bb88cb858-2lccf\" (UID: \"95afe811-a25f-4b89-9928-ee261693091b\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.006450 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2jgc\" (UniqueName: \"kubernetes.io/projected/6e6f9937-437a-4f97-a8ea-e486ec9a448c-kube-api-access-b2jgc\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-kxlc9\" (UID: \"6e6f9937-437a-4f97-a8ea-e486ec9a448c\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.013530 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.025993 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghrn7\" (UniqueName: \"kubernetes.io/projected/3b5de07d-ee5e-492a-9b9d-26de9837ca80-kube-api-access-ghrn7\") pod \"nova-operator-controller-manager-86d796d84d-5v5lr\" (UID: \"3b5de07d-ee5e-492a-9b9d-26de9837ca80\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.047925 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm2zl\" (UniqueName: \"kubernetes.io/projected/8a1443d1-4bea-481d-ad92-c92e7f5ac7cd-kube-api-access-tm2zl\") pod \"neutron-operator-controller-manager-66b7d6f598-xgq2g\" (UID: \"8a1443d1-4bea-481d-ad92-c92e7f5ac7cd\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.069829 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.071408 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.078793 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.080337 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zprkz\" (UniqueName: \"kubernetes.io/projected/fc6a7978-e9a3-4263-89d9-7ed03da8035f-kube-api-access-zprkz\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-2xzbc\" (UID: \"fc6a7978-e9a3-4263-89d9-7ed03da8035f\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.080407 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd4nr\" (UniqueName: \"kubernetes.io/projected/0fce00df-734a-4701-9f85-034b62ec61a9-kube-api-access-cd4nr\") pod \"octavia-operator-controller-manager-6fdc856c5d-k7jkc\" (UID: \"0fce00df-734a-4701-9f85-034b62ec61a9\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.080783 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-ngbgh" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.111326 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.114408 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.114768 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.123815 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-8xgj7" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.124110 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.133980 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.149192 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.150492 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.150701 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.167346 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-27lhh" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.172841 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.174347 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.175866 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-9rph7"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.177109 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.179524 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-2fx42" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.181301 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-mks92" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.182312 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tln26\" (UniqueName: \"kubernetes.io/projected/9d9e84d5-49ab-46aa-971e-cd920e2801b1-kube-api-access-tln26\") pod \"placement-operator-controller-manager-6dc664666c-74qt8\" (UID: \"9d9e84d5-49ab-46aa-971e-cd920e2801b1\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.182358 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zprkz\" (UniqueName: \"kubernetes.io/projected/fc6a7978-e9a3-4263-89d9-7ed03da8035f-kube-api-access-zprkz\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-2xzbc\" (UID: \"fc6a7978-e9a3-4263-89d9-7ed03da8035f\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.182401 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd4nr\" (UniqueName: \"kubernetes.io/projected/0fce00df-734a-4701-9f85-034b62ec61a9-kube-api-access-cd4nr\") pod \"octavia-operator-controller-manager-6fdc856c5d-k7jkc\" (UID: \"0fce00df-734a-4701-9f85-034b62ec61a9\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.192625 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.201983 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd4nr\" (UniqueName: \"kubernetes.io/projected/0fce00df-734a-4701-9f85-034b62ec61a9-kube-api-access-cd4nr\") pod \"octavia-operator-controller-manager-6fdc856c5d-k7jkc\" (UID: \"0fce00df-734a-4701-9f85-034b62ec61a9\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.215418 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zprkz\" (UniqueName: \"kubernetes.io/projected/fc6a7978-e9a3-4263-89d9-7ed03da8035f-kube-api-access-zprkz\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-2xzbc\" (UID: \"fc6a7978-e9a3-4263-89d9-7ed03da8035f\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.217257 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.253139 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.267808 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.272294 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.284947 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b0be97de-cd56-47c0-bd50-868a8fba50d6-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5\" (UID: \"b0be97de-cd56-47c0-bd50-868a8fba50d6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.285016 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26qv8\" (UniqueName: \"kubernetes.io/projected/66c8f0b0-71e8-4bf6-a28b-109bff76c35e-kube-api-access-26qv8\") pod \"swift-operator-controller-manager-799cb6ffd6-scbrc\" (UID: \"66c8f0b0-71e8-4bf6-a28b-109bff76c35e\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.285051 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tln26\" (UniqueName: \"kubernetes.io/projected/9d9e84d5-49ab-46aa-971e-cd920e2801b1-kube-api-access-tln26\") pod \"placement-operator-controller-manager-6dc664666c-74qt8\" (UID: \"9d9e84d5-49ab-46aa-971e-cd920e2801b1\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.285082 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g789p\" (UniqueName: \"kubernetes.io/projected/b0be97de-cd56-47c0-bd50-868a8fba50d6-kube-api-access-g789p\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5\" (UID: \"b0be97de-cd56-47c0-bd50-868a8fba50d6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.285112 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w67nl\" (UniqueName: \"kubernetes.io/projected/1d093c2b-22f5-4146-a0e1-a29ab68e1ab3-kube-api-access-w67nl\") pod \"telemetry-operator-controller-manager-7798859c74-d4b8t\" (UID: \"1d093c2b-22f5-4146-a0e1-a29ab68e1ab3\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.285163 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s76ks\" (UniqueName: \"kubernetes.io/projected/486bc897-b77f-41d9-8507-a667d595eb61-kube-api-access-s76ks\") pod \"test-operator-controller-manager-8464cf66df-9rph7\" (UID: \"486bc897-b77f-41d9-8507-a667d595eb61\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.292626 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-9rph7"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.309781 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.326750 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.327878 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tln26\" (UniqueName: \"kubernetes.io/projected/9d9e84d5-49ab-46aa-971e-cd920e2801b1-kube-api-access-tln26\") pod \"placement-operator-controller-manager-6dc664666c-74qt8\" (UID: \"9d9e84d5-49ab-46aa-971e-cd920e2801b1\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.333001 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.337724 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-9r9hj" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.343078 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.373047 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.374205 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.374774 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.375263 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.377815 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.378920 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-79dpk" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.380187 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-r9zct" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.387532 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s76ks\" (UniqueName: \"kubernetes.io/projected/486bc897-b77f-41d9-8507-a667d595eb61-kube-api-access-s76ks\") pod \"test-operator-controller-manager-8464cf66df-9rph7\" (UID: \"486bc897-b77f-41d9-8507-a667d595eb61\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.389763 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.392885 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b0be97de-cd56-47c0-bd50-868a8fba50d6-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5\" (UID: \"b0be97de-cd56-47c0-bd50-868a8fba50d6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.393078 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26qv8\" (UniqueName: \"kubernetes.io/projected/66c8f0b0-71e8-4bf6-a28b-109bff76c35e-kube-api-access-26qv8\") pod \"swift-operator-controller-manager-799cb6ffd6-scbrc\" (UID: \"66c8f0b0-71e8-4bf6-a28b-109bff76c35e\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" Nov 24 07:03:09 crc kubenswrapper[4799]: E1124 07:03:09.394083 4799 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 07:03:09 crc kubenswrapper[4799]: E1124 07:03:09.398808 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0be97de-cd56-47c0-bd50-868a8fba50d6-cert podName:b0be97de-cd56-47c0-bd50-868a8fba50d6 nodeName:}" failed. No retries permitted until 2025-11-24 07:03:09.898773344 +0000 UTC m=+935.554755818 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b0be97de-cd56-47c0-bd50-868a8fba50d6-cert") pod "openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" (UID: "b0be97de-cd56-47c0-bd50-868a8fba50d6") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.393185 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g789p\" (UniqueName: \"kubernetes.io/projected/b0be97de-cd56-47c0-bd50-868a8fba50d6-kube-api-access-g789p\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5\" (UID: \"b0be97de-cd56-47c0-bd50-868a8fba50d6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.399080 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztm4j\" (UniqueName: \"kubernetes.io/projected/9ecccf0c-10f4-43e6-b658-528ff9acea0f-kube-api-access-ztm4j\") pod \"watcher-operator-controller-manager-7cd4fb6f79-fghbt\" (UID: \"9ecccf0c-10f4-43e6-b658-528ff9acea0f\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.399564 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w67nl\" (UniqueName: \"kubernetes.io/projected/1d093c2b-22f5-4146-a0e1-a29ab68e1ab3-kube-api-access-w67nl\") pod \"telemetry-operator-controller-manager-7798859c74-d4b8t\" (UID: \"1d093c2b-22f5-4146-a0e1-a29ab68e1ab3\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.424174 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g789p\" (UniqueName: \"kubernetes.io/projected/b0be97de-cd56-47c0-bd50-868a8fba50d6-kube-api-access-g789p\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5\" (UID: \"b0be97de-cd56-47c0-bd50-868a8fba50d6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.429355 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w67nl\" (UniqueName: \"kubernetes.io/projected/1d093c2b-22f5-4146-a0e1-a29ab68e1ab3-kube-api-access-w67nl\") pod \"telemetry-operator-controller-manager-7798859c74-d4b8t\" (UID: \"1d093c2b-22f5-4146-a0e1-a29ab68e1ab3\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.435121 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.435582 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s76ks\" (UniqueName: \"kubernetes.io/projected/486bc897-b77f-41d9-8507-a667d595eb61-kube-api-access-s76ks\") pod \"test-operator-controller-manager-8464cf66df-9rph7\" (UID: \"486bc897-b77f-41d9-8507-a667d595eb61\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.455120 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26qv8\" (UniqueName: \"kubernetes.io/projected/66c8f0b0-71e8-4bf6-a28b-109bff76c35e-kube-api-access-26qv8\") pod \"swift-operator-controller-manager-799cb6ffd6-scbrc\" (UID: \"66c8f0b0-71e8-4bf6-a28b-109bff76c35e\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.502198 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztm4j\" (UniqueName: \"kubernetes.io/projected/9ecccf0c-10f4-43e6-b658-528ff9acea0f-kube-api-access-ztm4j\") pod \"watcher-operator-controller-manager-7cd4fb6f79-fghbt\" (UID: \"9ecccf0c-10f4-43e6-b658-528ff9acea0f\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.502271 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5c91af0a-fe03-45c5-ae6f-3ea056c17e31-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-px9nk\" (UID: \"5c91af0a-fe03-45c5-ae6f-3ea056c17e31\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.502339 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z29s8\" (UniqueName: \"kubernetes.io/projected/5c91af0a-fe03-45c5-ae6f-3ea056c17e31-kube-api-access-z29s8\") pod \"openstack-operator-controller-manager-6cb9dc54f8-px9nk\" (UID: \"5c91af0a-fe03-45c5-ae6f-3ea056c17e31\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.502382 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btfmb\" (UniqueName: \"kubernetes.io/projected/611f4120-8c1b-4745-a5d7-4e1dd52ec567-kube-api-access-btfmb\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-gv46d\" (UID: \"611f4120-8c1b-4745-a5d7-4e1dd52ec567\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.523575 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztm4j\" (UniqueName: \"kubernetes.io/projected/9ecccf0c-10f4-43e6-b658-528ff9acea0f-kube-api-access-ztm4j\") pod \"watcher-operator-controller-manager-7cd4fb6f79-fghbt\" (UID: \"9ecccf0c-10f4-43e6-b658-528ff9acea0f\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.553402 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.603794 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5c91af0a-fe03-45c5-ae6f-3ea056c17e31-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-px9nk\" (UID: \"5c91af0a-fe03-45c5-ae6f-3ea056c17e31\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.603878 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z29s8\" (UniqueName: \"kubernetes.io/projected/5c91af0a-fe03-45c5-ae6f-3ea056c17e31-kube-api-access-z29s8\") pod \"openstack-operator-controller-manager-6cb9dc54f8-px9nk\" (UID: \"5c91af0a-fe03-45c5-ae6f-3ea056c17e31\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.604011 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btfmb\" (UniqueName: \"kubernetes.io/projected/611f4120-8c1b-4745-a5d7-4e1dd52ec567-kube-api-access-btfmb\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-gv46d\" (UID: \"611f4120-8c1b-4745-a5d7-4e1dd52ec567\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d" Nov 24 07:03:09 crc kubenswrapper[4799]: E1124 07:03:09.604065 4799 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 07:03:09 crc kubenswrapper[4799]: E1124 07:03:09.604174 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c91af0a-fe03-45c5-ae6f-3ea056c17e31-cert podName:5c91af0a-fe03-45c5-ae6f-3ea056c17e31 nodeName:}" failed. No retries permitted until 2025-11-24 07:03:10.104145663 +0000 UTC m=+935.760128137 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5c91af0a-fe03-45c5-ae6f-3ea056c17e31-cert") pod "openstack-operator-controller-manager-6cb9dc54f8-px9nk" (UID: "5c91af0a-fe03-45c5-ae6f-3ea056c17e31") : secret "webhook-server-cert" not found Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.607533 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.625564 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btfmb\" (UniqueName: \"kubernetes.io/projected/611f4120-8c1b-4745-a5d7-4e1dd52ec567-kube-api-access-btfmb\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-gv46d\" (UID: \"611f4120-8c1b-4745-a5d7-4e1dd52ec567\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.627033 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z29s8\" (UniqueName: \"kubernetes.io/projected/5c91af0a-fe03-45c5-ae6f-3ea056c17e31-kube-api-access-z29s8\") pod \"openstack-operator-controller-manager-6cb9dc54f8-px9nk\" (UID: \"5c91af0a-fe03-45c5-ae6f-3ea056c17e31\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.633298 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.705148 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.722228 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.759399 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.820365 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.905681 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx"] Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.923076 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b0be97de-cd56-47c0-bd50-868a8fba50d6-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5\" (UID: \"b0be97de-cd56-47c0-bd50-868a8fba50d6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" Nov 24 07:03:09 crc kubenswrapper[4799]: I1124 07:03:09.933388 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b0be97de-cd56-47c0-bd50-868a8fba50d6-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5\" (UID: \"b0be97de-cd56-47c0-bd50-868a8fba50d6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.017404 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.126606 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5c91af0a-fe03-45c5-ae6f-3ea056c17e31-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-px9nk\" (UID: \"5c91af0a-fe03-45c5-ae6f-3ea056c17e31\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.131080 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5c91af0a-fe03-45c5-ae6f-3ea056c17e31-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-px9nk\" (UID: \"5c91af0a-fe03-45c5-ae6f-3ea056c17e31\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.163507 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.179614 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.270623 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.303426 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.312439 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9"] Nov 24 07:03:10 crc kubenswrapper[4799]: W1124 07:03:10.315658 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf65b87f_5fe7_4ade_873b_e4d7040df219.slice/crio-b735bc88f001861d0bb2cf7b31e9f7cc4643e2f672b67da6bc89a7af1846cdc3 WatchSource:0}: Error finding container b735bc88f001861d0bb2cf7b31e9f7cc4643e2f672b67da6bc89a7af1846cdc3: Status 404 returned error can't find the container with id b735bc88f001861d0bb2cf7b31e9f7cc4643e2f672b67da6bc89a7af1846cdc3 Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.324902 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.339196 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.466244 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt" event={"ID":"f1284cfd-b954-4b2d-8e3c-2e57bb52ca27","Type":"ContainerStarted","Data":"6818f656351a1d84d4cb2c13734a8af43fe173c2eee2d920843b9c3d6c3d10ed"} Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.471874 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl" event={"ID":"4866479c-d0b9-4366-8ab4-f041573d0d35","Type":"ContainerStarted","Data":"de4c44d3f36293a4c65fb3922bf51b66d9ba4d470ea212da621def412533870b"} Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.473562 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9" event={"ID":"6e6f9937-437a-4f97-a8ea-e486ec9a448c","Type":"ContainerStarted","Data":"591f1a7f1be991381fb5ee3201477b51c7dd9af6ab4edb6642baff450fa7fe54"} Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.476420 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx" event={"ID":"3fec8392-f37a-4f4a-bdec-696bf7b02cd1","Type":"ContainerStarted","Data":"f1572eb9d88921e0c997f51f32d16a0309a7f1f37ea1a78353110a1f3e57abea"} Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.478144 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd" event={"ID":"af65b87f-5fe7-4ade-873b-e4d7040df219","Type":"ContainerStarted","Data":"b735bc88f001861d0bb2cf7b31e9f7cc4643e2f672b67da6bc89a7af1846cdc3"} Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.480477 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2" event={"ID":"9f756dcc-daa5-4ed3-a4af-3bbf3c9f186c","Type":"ContainerStarted","Data":"318e145a46b8b33d1161a404483ca8e7d99e09a15c029cda13e4cccc830dc23b"} Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.484068 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld" event={"ID":"52251a10-6203-43da-af96-94f78f0edacc","Type":"ContainerStarted","Data":"1e55db7184fd180b04b39508d11d8f514a63d3352ca8cdf38d680efcdd263c40"} Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.488272 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr" event={"ID":"3b5de07d-ee5e-492a-9b9d-26de9837ca80","Type":"ContainerStarted","Data":"8994aa889633869639fdc534432c266fd594a405db97e2eca37e38ed9141bddd"} Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.495503 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.526477 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc"] Nov 24 07:03:10 crc kubenswrapper[4799]: W1124 07:03:10.534962 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fce00df_734a_4701_9f85_034b62ec61a9.slice/crio-3d113843a4c7707fdca50e2fee62ea7a100174362c9e85150dc21b35434c2be7 WatchSource:0}: Error finding container 3d113843a4c7707fdca50e2fee62ea7a100174362c9e85150dc21b35434c2be7: Status 404 returned error can't find the container with id 3d113843a4c7707fdca50e2fee62ea7a100174362c9e85150dc21b35434c2be7 Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.681840 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.708645 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.741896 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.761136 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn"] Nov 24 07:03:10 crc kubenswrapper[4799]: W1124 07:03:10.764813 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6a55aa5_3e1c_49e9_bd56_48f8497c1b49.slice/crio-ebca7a5c6bdbf32d62058f3956499ed8fed679d45f48758f51eeaa2d0c6bddb1 WatchSource:0}: Error finding container ebca7a5c6bdbf32d62058f3956499ed8fed679d45f48758f51eeaa2d0c6bddb1: Status 404 returned error can't find the container with id ebca7a5c6bdbf32d62058f3956499ed8fed679d45f48758f51eeaa2d0c6bddb1 Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.771376 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf"] Nov 24 07:03:10 crc kubenswrapper[4799]: W1124 07:03:10.771554 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a1443d1_4bea_481d_ad92_c92e7f5ac7cd.slice/crio-02c1b0983bba7e9ab9d0b777d7c1e5456b93a89142da2a8ee8033f95efc4f6c5 WatchSource:0}: Error finding container 02c1b0983bba7e9ab9d0b777d7c1e5456b93a89142da2a8ee8033f95efc4f6c5: Status 404 returned error can't find the container with id 02c1b0983bba7e9ab9d0b777d7c1e5456b93a89142da2a8ee8033f95efc4f6c5 Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.777259 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc"] Nov 24 07:03:10 crc kubenswrapper[4799]: E1124 07:03:10.778264 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tm2zl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-66b7d6f598-xgq2g_openstack-operators(8a1443d1-4bea-481d-ad92-c92e7f5ac7cd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.785750 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.790492 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5"] Nov 24 07:03:10 crc kubenswrapper[4799]: W1124 07:03:10.791047 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0be97de_cd56_47c0_bd50_868a8fba50d6.slice/crio-5f8f5aefe78c5e93fc9ae1e7e1f89b4aa23a60b8b55f5dcf2ce0e23a3b4e6300 WatchSource:0}: Error finding container 5f8f5aefe78c5e93fc9ae1e7e1f89b4aa23a60b8b55f5dcf2ce0e23a3b4e6300: Status 404 returned error can't find the container with id 5f8f5aefe78c5e93fc9ae1e7e1f89b4aa23a60b8b55f5dcf2ce0e23a3b4e6300 Nov 24 07:03:10 crc kubenswrapper[4799]: E1124 07:03:10.799783 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zprkz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-5bdf4f7f7f-2xzbc_openstack-operators(fc6a7978-e9a3-4263-89d9-7ed03da8035f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:03:10 crc kubenswrapper[4799]: E1124 07:03:10.799525 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:7dbadf7b98f2f305f9f1382f55a084c8ca404f4263f76b28e56bd0dc437e2192,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:c8101c77a82eae4407e41e1fd766dfc6e1b7f9ed1679e3efb6f91ff97a1557b2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:eb9743b21bbadca6f7cb9ac4fc46b5d58c51c674073c7e1121f4474a71304071,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:3d81f839b98c2e2a5bf0da79f2f9a92dff7d0a3c5a830b0e95c89dad8cf98a6a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:d19ac99249b47dd8ea16cd6aaa5756346aa8a2f119ee50819c15c5366efb417d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:8536169e5537fe6c330eba814248abdcf39cdd8f7e7336034d74e6fda9544050,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:4f1fa337760e82bfd67cdd142a97c121146dd7e621daac161940dd5e4ddb80dc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:3613b345d5baed98effd906f8b0242d863e14c97078ea473ef01fe1b0afc46f3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:d375d370be5ead0dac71109af644849e5795f535f9ad8eeacea261d77ae6f140,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:9f9f367ed4c85efb16c3a74a4bb707ff0db271d7bc5abc70a71e984b55f43003,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:b73ad22b4955b06d584bce81742556d8c0c7828c495494f8ea7c99391c61b70f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:aa1d3aaf6b394621ed4089a98e0a82b763f467e8b5c5db772f9fdf99fc86e333,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:d6661053141b6df421288a7c9968a155ab82e478c1d75ab41f2cebe2f0ca02d2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:ce2d63258cb4e7d0d1c07234de6889c5434464190906798019311a1c7cf6387f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:0485ef9e5b4437f7cd2ba54034a87722ce4669ee86b3773c6b0c037ed8000e91,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api@sha256:962c004551d0503779364b767b9bf0cecdf78dbba8809b2ca8b073f58e1f4e5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor@sha256:0ebf4c465fb6cc7dad9e6cb2da0ff54874c9acbcb40d62234a629ec2c12cdd62,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:ff0c553ceeb2e0f44b010e37dc6d0db8a251797b88e56468b7cf7f05253e4232,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:624f553f073af7493d34828b074adc9981cce403edd8e71482c7307008479fd9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:e3874936a518c8560339db8f840fc5461885819f6050b5de8d3ab9199bea5094,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:1cea25f1d2a45affc80c46fb9d427749d3f06b61590ac6070a2910e3ec8a4e5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:e36d5b9a65194f12f7b01c6422ba3ed52a687fd1695fbb21f4986c67d9f9317f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:8b21bec527d54cd766e277889df6bcccd2baeaa946274606b986c0c3b7ca689f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:45aceca77f8fcf61127f0da650bdfdf11ede9b0944c78b63fab819d03283f96b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:709ac58998927dd61786821ae1e63343fd97ccf5763aac5edb4583eea9401d22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:867d4ef7c21f75e6030a685b5762ab4d84b671316ed6b98d75200076e93342cd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:2b90da93550b99d2fcfa95bd819f3363aa68346a416f8dc7baac3e9c5f487761,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:6f86db36d668348be8c5b46dcda8b1fa23d34bfdc07164fbcbe7a6327fb4de24,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:8cde52cef8795d1c91983b100d86541c7718160ec260fe0f97b96add4c2c8ee8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:a9583cb3baf440d2358ef041373833afbeae60da8159dd031502379901141620,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:835ebed082fe1c45bd799d1d5357595ce63efeb05ca876f26b08443facb9c164,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:011d682241db724bc40736c9b54d2ea450ea7e6be095b1ff5fa28c8007466775,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:2025da90cff8f563deb08bee71efe16d4078edc2a767b2e225cca5c77f1aa2f9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:26bd7b0bd6070856aefef6fe754c547d55c056396ea30d879d34c2d49b5a1d29,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:ff46cd5e0e13d105c4629e78c2734a50835f06b6a1e31da9e0462981d10c4be3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:5b4fd0c2b76fa5539f74687b11c5882d77bd31352452322b37ff51fa18f12a61,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:5e03376bd895346dc8f627ca15ded942526ed8b5e92872f453ce272e694d18d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:36a0fb31978aee0ded2483de311631e64a644d0b0685b5b055f65ede7eb8e8a2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:5f6045841aff0fde6f684a34cdf49f8dc7b2c3bcbdeab201f1058971e0c5f79e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:448f4e1b740c30936e340bd6e8534d78c83357bf373a4223950aa64d3484f007,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:b68e3615af8a0eb0ef6bf9ceeef59540a6f4a9a85f6078a3620be115c73a7db8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:7eae01cf60383e523c9cd94d158a9162120a7370829a1dad20fdea6b0fd660bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:28cc10501788081eb61b5a1af35546191a92741f4f109df54c74e2b19439d0f9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:9a616e37acfd120612f78043237a8541266ba34883833c9beb43f3da313661ad,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:6b1be6cd94a0942259bca5d5d2c30cc7de4a33276b61f8ae3940226772106256,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:02d2c22d15401574941fbe057095442dee0d6f7a0a9341de35d25e6a12a3fe4b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:fc3b3a36b74fd653946723c54b208072d52200635850b531e9d595a7aaea5a01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:7850ccbff320bf9a1c9c769c1c70777eb97117dd8cd5ae4435be9b4622cf807a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:397dac7e39cf40d14a986e6ec4a60fb698ca35c197d0db315b1318514cc6d1d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:1c95142a36276686e720f86423ee171dc9adcc1e89879f627545b7c906ccd9bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:e331a8fde6638e5ba154c4f0b38772a9a424f60656f2777245975fb1fa02f07d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:b6e1e8a249d36ef36c6ac4170af1e043dda1ccc0f9672832d3ff151bf3533076,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:cd3cf7a34053e850b4d4f9f4ea4c74953a54a42fd18e47d7c01d44a88923e925,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:aee28476344fc0cc148fbe97daf9b1bfcedc22001550bba4bdc4e84be7b6989d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:cfa0b92c976603ee2a937d34013a238fcd8aa75f998e50642e33489f14124633,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:73c2f2d6eecf88acf4e45b133c8373d9bb006b530e0aff0b28f3b7420620a874,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:927b405cc04abe5ff716186e8d35e2dc5fad1c8430194659ee6617d74e4e055d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:6154d7cebd7c339afa5b86330262156171743aa5b79c2b78f9a2f378005ed8fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:e2db2f4af8d3d0be7868c6efef0189f3a2c74a8f96ae10e3f991cdf83feaef29,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:c773629df257726a6d3cacc24a6e4df0babcd7d37df04e6d14676a8da028b9c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:776211111e2e6493706dbc49a3ba44f31d1b947919313ed3a0f35810e304ec52,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:0a98e8f5c83522ca6c8e40c5e9561f6628d2d5e69f0e8a64279c541c989d3d8b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:7cccf24ad0a152f90ca39893064f48a1656950ee8142685a5d482c71f0bdc9f5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:af46761060c7987e1dee5f14c06d85b46f12ad8e09c83d4246ab4e3a65dfda3e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:05450b48f6b5352b2686a26e933e8727748edae2ae9652d9164b7d7a1817c55a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:fc9c99eeef91523482bd8f92661b393287e1f2a24ad2ba9e33191f8de9af74cf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:3e4ecc02b4b5e0860482a93599ba9ca598c5ce26c093c46e701f96fe51acb208,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:2346037e064861c7892690d2e8b3e1eea1a26ce3c3a11fda0b41301965bc828c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:c26c3ff9cabe3593ceb10006e782bf9391ac14785768ce9eec4f938c2d3cf228,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:daa45220bb1c47922d0917aa8fe423bb82b03a01429f1c9e37635e701e352d71,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:a80a074e227d3238bb6f285788a9e886ae7a5909ccbc5c19c93c369bdfe5b3b8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:58ac66ca1be01fe0157977bd79a26cde4d0de153edfaf4162367c924826b2ef4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:99a63770d80cc7c3afa1118b400972fb0e6bff5284a2eae781b12582ad79c29c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:9ee4d84529394afcd860f1a1186484560f02f08c15c37cac42a22473b7116d5f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:ea15fadda7b0439ec637edfaf6ea5dbf3e35fb3be012c7c5a31e722c90becb11,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g789p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5_openstack-operators(b0be97de-cd56-47c0-bd50-868a8fba50d6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.876094 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-9rph7"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.881960 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.888327 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8"] Nov 24 07:03:10 crc kubenswrapper[4799]: W1124 07:03:10.889360 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d9e84d5_49ab_46aa_971e_cd920e2801b1.slice/crio-53e59893b1f375ac11cb652043f65c59607135b1b6237677b59e3c1a4e04d3f7 WatchSource:0}: Error finding container 53e59893b1f375ac11cb652043f65c59607135b1b6237677b59e3c1a4e04d3f7: Status 404 returned error can't find the container with id 53e59893b1f375ac11cb652043f65c59607135b1b6237677b59e3c1a4e04d3f7 Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.892786 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d"] Nov 24 07:03:10 crc kubenswrapper[4799]: I1124 07:03:10.896928 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t"] Nov 24 07:03:10 crc kubenswrapper[4799]: W1124 07:03:10.917787 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod486bc897_b77f_41d9_8507_a667d595eb61.slice/crio-b64ed84d69be64549db01f0dd2c40203d24620690a355e52882e16f5a461b0c3 WatchSource:0}: Error finding container b64ed84d69be64549db01f0dd2c40203d24620690a355e52882e16f5a461b0c3: Status 404 returned error can't find the container with id b64ed84d69be64549db01f0dd2c40203d24620690a355e52882e16f5a461b0c3 Nov 24 07:03:10 crc kubenswrapper[4799]: W1124 07:03:10.919183 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod611f4120_8c1b_4745_a5d7_4e1dd52ec567.slice/crio-47e270020d59a99462e015089a23e73547871b43e5158a5af688d30044bf448a WatchSource:0}: Error finding container 47e270020d59a99462e015089a23e73547871b43e5158a5af688d30044bf448a: Status 404 returned error can't find the container with id 47e270020d59a99462e015089a23e73547871b43e5158a5af688d30044bf448a Nov 24 07:03:10 crc kubenswrapper[4799]: W1124 07:03:10.920190 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d093c2b_22f5_4146_a0e1_a29ab68e1ab3.slice/crio-8183cf15e50d2e72a6510f8969ed99fcdf3409b28a4ef6532e3f202cf2045eae WatchSource:0}: Error finding container 8183cf15e50d2e72a6510f8969ed99fcdf3409b28a4ef6532e3f202cf2045eae: Status 404 returned error can't find the container with id 8183cf15e50d2e72a6510f8969ed99fcdf3409b28a4ef6532e3f202cf2045eae Nov 24 07:03:10 crc kubenswrapper[4799]: W1124 07:03:10.921281 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66c8f0b0_71e8_4bf6_a28b_109bff76c35e.slice/crio-86a7aa54bb5e6f2b159aab213caa3946297e3a2a185a8e79d191adac7729865d WatchSource:0}: Error finding container 86a7aa54bb5e6f2b159aab213caa3946297e3a2a185a8e79d191adac7729865d: Status 404 returned error can't find the container with id 86a7aa54bb5e6f2b159aab213caa3946297e3a2a185a8e79d191adac7729865d Nov 24 07:03:10 crc kubenswrapper[4799]: E1124 07:03:10.932753 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w67nl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7798859c74-d4b8t_openstack-operators(1d093c2b-22f5-4146-a0e1-a29ab68e1ab3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:03:10 crc kubenswrapper[4799]: E1124 07:03:10.932936 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s76ks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-8464cf66df-9rph7_openstack-operators(486bc897-b77f-41d9-8507-a667d595eb61): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:03:10 crc kubenswrapper[4799]: E1124 07:03:10.933409 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-26qv8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-799cb6ffd6-scbrc_openstack-operators(66c8f0b0-71e8-4bf6-a28b-109bff76c35e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:03:10 crc kubenswrapper[4799]: E1124 07:03:10.936459 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-btfmb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-gv46d_openstack-operators(611f4120-8c1b-4745-a5d7-4e1dd52ec567): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 07:03:10 crc kubenswrapper[4799]: E1124 07:03:10.937924 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d" podUID="611f4120-8c1b-4745-a5d7-4e1dd52ec567" Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.021984 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" podUID="8a1443d1-4bea-481d-ad92-c92e7f5ac7cd" Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.027150 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" podUID="fc6a7978-e9a3-4263-89d9-7ed03da8035f" Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.138423 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" podUID="486bc897-b77f-41d9-8507-a667d595eb61" Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.161769 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" podUID="b0be97de-cd56-47c0-bd50-868a8fba50d6" Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.226207 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" podUID="66c8f0b0-71e8-4bf6-a28b-109bff76c35e" Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.264903 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" podUID="1d093c2b-22f5-4146-a0e1-a29ab68e1ab3" Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.506123 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8" event={"ID":"9d9e84d5-49ab-46aa-971e-cd920e2801b1","Type":"ContainerStarted","Data":"53e59893b1f375ac11cb652043f65c59607135b1b6237677b59e3c1a4e04d3f7"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.508010 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" event={"ID":"9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c","Type":"ContainerStarted","Data":"b15ecc379ced05334dd5858dcda9e39500027b456abca4a2fa3118a25926bc9e"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.518242 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" event={"ID":"66c8f0b0-71e8-4bf6-a28b-109bff76c35e","Type":"ContainerStarted","Data":"8e301dfcc2308339ea97a4b82f564fdd0610e66bb703c559e2b837466915d04b"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.518302 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" event={"ID":"66c8f0b0-71e8-4bf6-a28b-109bff76c35e","Type":"ContainerStarted","Data":"86a7aa54bb5e6f2b159aab213caa3946297e3a2a185a8e79d191adac7729865d"} Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.541354 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" podUID="66c8f0b0-71e8-4bf6-a28b-109bff76c35e" Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.548054 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" event={"ID":"8a1443d1-4bea-481d-ad92-c92e7f5ac7cd","Type":"ContainerStarted","Data":"41d1a6ffcf51c204302b18a209ba1faf0b60e274a961a5aea8c9945b902cfdb7"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.548111 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" event={"ID":"8a1443d1-4bea-481d-ad92-c92e7f5ac7cd","Type":"ContainerStarted","Data":"02c1b0983bba7e9ab9d0b777d7c1e5456b93a89142da2a8ee8033f95efc4f6c5"} Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.551187 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" podUID="8a1443d1-4bea-481d-ad92-c92e7f5ac7cd" Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.566589 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" event={"ID":"5c91af0a-fe03-45c5-ae6f-3ea056c17e31","Type":"ContainerStarted","Data":"7979ea82fcdf471a9f2f164d6ba3a7ec5dc447ba49a8d85b25113571b0332d30"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.566645 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" event={"ID":"5c91af0a-fe03-45c5-ae6f-3ea056c17e31","Type":"ContainerStarted","Data":"98b50f7ece0e9f6b376b83848515df1ed0ae77a232ec68bf567da6ac276635a9"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.566657 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" event={"ID":"5c91af0a-fe03-45c5-ae6f-3ea056c17e31","Type":"ContainerStarted","Data":"807ede014e8c2fa214fa9e473ead9cdbcd6a0ae2c243022ffd85e185ebc5b9bc"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.567044 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.573028 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" event={"ID":"fc6a7978-e9a3-4263-89d9-7ed03da8035f","Type":"ContainerStarted","Data":"96c6273ce4124199bf405466d67279fc15ab169b76b39fe7a1710865be5bc79b"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.578639 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" event={"ID":"fc6a7978-e9a3-4263-89d9-7ed03da8035f","Type":"ContainerStarted","Data":"c71d7f16fc4b48178dd8c3f2a1a68fe7a4f6041a9d9bd4a4aabd4d7c8f1e6a42"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.589215 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf" event={"ID":"95afe811-a25f-4b89-9928-ee261693091b","Type":"ContainerStarted","Data":"62e833b3ab586ee6f161cf703f304f43450c013489b0629ddd50235f6049515d"} Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.589494 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" podUID="fc6a7978-e9a3-4263-89d9-7ed03da8035f" Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.636255 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" podStartSLOduration=2.636227401 podStartE2EDuration="2.636227401s" podCreationTimestamp="2025-11-24 07:03:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:03:11.623176972 +0000 UTC m=+937.279159446" watchObservedRunningTime="2025-11-24 07:03:11.636227401 +0000 UTC m=+937.292209875" Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.642507 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" event={"ID":"1d093c2b-22f5-4146-a0e1-a29ab68e1ab3","Type":"ContainerStarted","Data":"e5c9ebdae7731259121ee588461d2d83ea38dc3584532cc046a92ad79413da01"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.642564 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" event={"ID":"1d093c2b-22f5-4146-a0e1-a29ab68e1ab3","Type":"ContainerStarted","Data":"8183cf15e50d2e72a6510f8969ed99fcdf3409b28a4ef6532e3f202cf2045eae"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.645032 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw" event={"ID":"32b50dc4-f35a-419b-9a18-9ff05f4c826c","Type":"ContainerStarted","Data":"a875d9c6c1ad8f5057044a9db6029531f2b6ff3b6dc4f8ff4108947e5e97c4a5"} Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.649445 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" podUID="1d093c2b-22f5-4146-a0e1-a29ab68e1ab3" Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.651574 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d" event={"ID":"611f4120-8c1b-4745-a5d7-4e1dd52ec567","Type":"ContainerStarted","Data":"47e270020d59a99462e015089a23e73547871b43e5158a5af688d30044bf448a"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.659222 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn" event={"ID":"d6a55aa5-3e1c-49e9-bd56-48f8497c1b49","Type":"ContainerStarted","Data":"ebca7a5c6bdbf32d62058f3956499ed8fed679d45f48758f51eeaa2d0c6bddb1"} Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.659252 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d" podUID="611f4120-8c1b-4745-a5d7-4e1dd52ec567" Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.663197 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" event={"ID":"b0be97de-cd56-47c0-bd50-868a8fba50d6","Type":"ContainerStarted","Data":"94336051483b1b814bbf95f1aa9b5d8104542743f14d63ba2bb7d689fd6af518"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.663236 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" event={"ID":"b0be97de-cd56-47c0-bd50-868a8fba50d6","Type":"ContainerStarted","Data":"5f8f5aefe78c5e93fc9ae1e7e1f89b4aa23a60b8b55f5dcf2ce0e23a3b4e6300"} Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.665285 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" podUID="b0be97de-cd56-47c0-bd50-868a8fba50d6" Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.673351 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" event={"ID":"486bc897-b77f-41d9-8507-a667d595eb61","Type":"ContainerStarted","Data":"ad6597f72a1c982bc50408611e811c71ce3c3a9d63f987a714912359f4a4d34f"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.673406 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" event={"ID":"486bc897-b77f-41d9-8507-a667d595eb61","Type":"ContainerStarted","Data":"b64ed84d69be64549db01f0dd2c40203d24620690a355e52882e16f5a461b0c3"} Nov 24 07:03:11 crc kubenswrapper[4799]: E1124 07:03:11.675953 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" podUID="486bc897-b77f-41d9-8507-a667d595eb61" Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.676787 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt" event={"ID":"9ecccf0c-10f4-43e6-b658-528ff9acea0f","Type":"ContainerStarted","Data":"f65ad9dc00ab7c0cff12d09de5773c352888dd3a3a708152e1ce4fd7170a19e8"} Nov 24 07:03:11 crc kubenswrapper[4799]: I1124 07:03:11.680929 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc" event={"ID":"0fce00df-734a-4701-9f85-034b62ec61a9","Type":"ContainerStarted","Data":"3d113843a4c7707fdca50e2fee62ea7a100174362c9e85150dc21b35434c2be7"} Nov 24 07:03:12 crc kubenswrapper[4799]: E1124 07:03:12.695731 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" podUID="fc6a7978-e9a3-4263-89d9-7ed03da8035f" Nov 24 07:03:12 crc kubenswrapper[4799]: E1124 07:03:12.697586 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" podUID="8a1443d1-4bea-481d-ad92-c92e7f5ac7cd" Nov 24 07:03:12 crc kubenswrapper[4799]: E1124 07:03:12.697660 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" podUID="66c8f0b0-71e8-4bf6-a28b-109bff76c35e" Nov 24 07:03:12 crc kubenswrapper[4799]: E1124 07:03:12.697705 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" podUID="b0be97de-cd56-47c0-bd50-868a8fba50d6" Nov 24 07:03:12 crc kubenswrapper[4799]: E1124 07:03:12.697736 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" podUID="1d093c2b-22f5-4146-a0e1-a29ab68e1ab3" Nov 24 07:03:12 crc kubenswrapper[4799]: E1124 07:03:12.699075 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" podUID="486bc897-b77f-41d9-8507-a667d595eb61" Nov 24 07:03:12 crc kubenswrapper[4799]: E1124 07:03:12.699602 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d" podUID="611f4120-8c1b-4745-a5d7-4e1dd52ec567" Nov 24 07:03:20 crc kubenswrapper[4799]: I1124 07:03:20.175710 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-px9nk" Nov 24 07:03:20 crc kubenswrapper[4799]: I1124 07:03:20.401409 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:03:20 crc kubenswrapper[4799]: I1124 07:03:20.401491 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:03:20 crc kubenswrapper[4799]: I1124 07:03:20.401546 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:03:20 crc kubenswrapper[4799]: I1124 07:03:20.402295 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"985590a83c847e8770acbc79642e467b641116d42bb23156312a6fcbb0de12d9"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:03:20 crc kubenswrapper[4799]: I1124 07:03:20.402405 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://985590a83c847e8770acbc79642e467b641116d42bb23156312a6fcbb0de12d9" gracePeriod=600 Nov 24 07:03:21 crc kubenswrapper[4799]: I1124 07:03:21.772458 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="985590a83c847e8770acbc79642e467b641116d42bb23156312a6fcbb0de12d9" exitCode=0 Nov 24 07:03:21 crc kubenswrapper[4799]: I1124 07:03:21.772547 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"985590a83c847e8770acbc79642e467b641116d42bb23156312a6fcbb0de12d9"} Nov 24 07:03:21 crc kubenswrapper[4799]: I1124 07:03:21.772983 4799 scope.go:117] "RemoveContainer" containerID="d29a0547b29f459c7d7db9037c62a8945e7f6c626531ca09a0a9fd2c39286a5a" Nov 24 07:03:22 crc kubenswrapper[4799]: I1124 07:03:22.926556 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd" event={"ID":"af65b87f-5fe7-4ade-873b-e4d7040df219","Type":"ContainerStarted","Data":"c397cfd5b1d9d048f00447e5bfce27e2ff0ff0e781a8c79ea4c700bcaa21b320"} Nov 24 07:03:22 crc kubenswrapper[4799]: I1124 07:03:22.951970 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr" event={"ID":"3b5de07d-ee5e-492a-9b9d-26de9837ca80","Type":"ContainerStarted","Data":"88e0df5713d47ec75dddad2199a59ca7218306fce2ecbfab3b73750eb4bdd04a"} Nov 24 07:03:23 crc kubenswrapper[4799]: I1124 07:03:22.999249 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl" event={"ID":"4866479c-d0b9-4366-8ab4-f041573d0d35","Type":"ContainerStarted","Data":"51612410f72c2eae916b7f5b237c2359d754c27de451345910ea6995cdcb2335"} Nov 24 07:03:23 crc kubenswrapper[4799]: I1124 07:03:23.020336 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx" event={"ID":"3fec8392-f37a-4f4a-bdec-696bf7b02cd1","Type":"ContainerStarted","Data":"f4adf2975a6815892e04d3926f4abaeb1c72ddc1ffad0e35d00d3765d0704f6a"} Nov 24 07:03:23 crc kubenswrapper[4799]: I1124 07:03:23.037254 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" event={"ID":"9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c","Type":"ContainerStarted","Data":"1c1de31254101f877fe961a46945dea24e96108ceb2b5dfff33234672b346cd6"} Nov 24 07:03:23 crc kubenswrapper[4799]: I1124 07:03:23.063416 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2" event={"ID":"9f756dcc-daa5-4ed3-a4af-3bbf3c9f186c","Type":"ContainerStarted","Data":"f4ba51827f4aa986e51c4c179fa634054f5fd46c75bb272a9f4edfa2fce67009"} Nov 24 07:03:23 crc kubenswrapper[4799]: I1124 07:03:23.119071 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt" event={"ID":"f1284cfd-b954-4b2d-8e3c-2e57bb52ca27","Type":"ContainerStarted","Data":"baa175c0d427fc56ab8e93be1b11d1dca2c0d81be7fe9e082008afd46d9ee79d"} Nov 24 07:03:23 crc kubenswrapper[4799]: I1124 07:03:23.130866 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw" event={"ID":"32b50dc4-f35a-419b-9a18-9ff05f4c826c","Type":"ContainerStarted","Data":"c6ad3dc76de1807bc3bc87a6ef06476d7b75579eeb526ac09eaed4f3ba8e045d"} Nov 24 07:03:23 crc kubenswrapper[4799]: I1124 07:03:23.153001 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf" event={"ID":"95afe811-a25f-4b89-9928-ee261693091b","Type":"ContainerStarted","Data":"d2bf34ef8195cf3a27051e25a9d2a9b3d9540b866325806fa2460c4bbab8ff96"} Nov 24 07:03:23 crc kubenswrapper[4799]: I1124 07:03:23.176298 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"9829ac623e13e600dbb23e96d6ce6d0e396075d576f57770ed87771c2d8517cf"} Nov 24 07:03:23 crc kubenswrapper[4799]: I1124 07:03:23.199213 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9" event={"ID":"6e6f9937-437a-4f97-a8ea-e486ec9a448c","Type":"ContainerStarted","Data":"78a4b296fd2ca56b05c2cf2a6a94ca30be8b2a78851821e3f30f5738da8f56c9"} Nov 24 07:03:23 crc kubenswrapper[4799]: I1124 07:03:23.208322 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn" event={"ID":"d6a55aa5-3e1c-49e9-bd56-48f8497c1b49","Type":"ContainerStarted","Data":"f201e72b316997e3f50f3a8933c0837bc8fc7a73d74e76ccc96c0e7d338e5047"} Nov 24 07:03:23 crc kubenswrapper[4799]: I1124 07:03:23.209686 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld" event={"ID":"52251a10-6203-43da-af96-94f78f0edacc","Type":"ContainerStarted","Data":"3fa58cd8d50503a67a7284c7f23c4fb42a3f3d3d3b111c8ea63e3925df5b19b5"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.218411 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld" event={"ID":"52251a10-6203-43da-af96-94f78f0edacc","Type":"ContainerStarted","Data":"e5f452d55a68c11ccc914da5b5f75576c009cd42ea7cd5ed409794c5665c4528"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.221080 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.224197 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt" event={"ID":"f1284cfd-b954-4b2d-8e3c-2e57bb52ca27","Type":"ContainerStarted","Data":"2a2f0836306522705a26c419455286d51ed59a1118ea46ac2094c824c1eb8b24"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.224385 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.225864 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc" event={"ID":"0fce00df-734a-4701-9f85-034b62ec61a9","Type":"ContainerStarted","Data":"31fb741b9d1869aacb35ee9e420d941facd27d67b56a09e0f20908846ac037b4"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.225911 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc" event={"ID":"0fce00df-734a-4701-9f85-034b62ec61a9","Type":"ContainerStarted","Data":"61d6902d572ff4bf4e5f01013cbb4a062acbd6063ab3d72edaa55d1c0e50d3b3"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.227321 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd" event={"ID":"af65b87f-5fe7-4ade-873b-e4d7040df219","Type":"ContainerStarted","Data":"4231ec8f98e81490d76a8f839f8c00cbafc07bd5d9f02a4f67a8f77be8bc1b23"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.228088 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.230084 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw" event={"ID":"32b50dc4-f35a-419b-9a18-9ff05f4c826c","Type":"ContainerStarted","Data":"8c1e9a05b6f7a6eaa302f41de6ab41f902baf6133c747a2f491b00fd5559a71f"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.230273 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.232247 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9" event={"ID":"6e6f9937-437a-4f97-a8ea-e486ec9a448c","Type":"ContainerStarted","Data":"fdba87bc276ac6ba7804041ed34cdd172ffa93ec9761f2201d3fc6cd502aa8b5"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.232446 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.234259 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8" event={"ID":"9d9e84d5-49ab-46aa-971e-cd920e2801b1","Type":"ContainerStarted","Data":"005c6026443943540c4487863a07ad5abfc8e914a067331fffdcf3aa10450f61"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.234290 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8" event={"ID":"9d9e84d5-49ab-46aa-971e-cd920e2801b1","Type":"ContainerStarted","Data":"5d23dc92f8cdedc59ac333e3cac41a5fdfde037ab5b07bfd7f9bb00197f3e763"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.234342 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.236171 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" event={"ID":"9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c","Type":"ContainerStarted","Data":"781d1a6a72bc7902bd48e0483b14fe5f46f567af739e1c2b56d69dbe3cad44e3"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.236316 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.237962 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx" event={"ID":"3fec8392-f37a-4f4a-bdec-696bf7b02cd1","Type":"ContainerStarted","Data":"e74a95195e58ad616036ec01d16fb84f2ee7265e0f09fa7309eca59b50636e2e"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.238257 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.240179 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2" event={"ID":"9f756dcc-daa5-4ed3-a4af-3bbf3c9f186c","Type":"ContainerStarted","Data":"90db524bbca1ae81d970474ecfd2e3cc56628de0f146fe9a9d0a92eec8d107af"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.240454 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.242434 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt" event={"ID":"9ecccf0c-10f4-43e6-b658-528ff9acea0f","Type":"ContainerStarted","Data":"cc9f35ec83a1e0ad6b1e7507b30d2814cb48c45303074ca8a8caccc9fa6ea225"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.242637 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.242794 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt" event={"ID":"9ecccf0c-10f4-43e6-b658-528ff9acea0f","Type":"ContainerStarted","Data":"8e51ef390d5ee4d719c1b06129f0bbf5e90389f12c8f910cdf1805cc9f563d4d"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.244600 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf" event={"ID":"95afe811-a25f-4b89-9928-ee261693091b","Type":"ContainerStarted","Data":"5d9cf66c121dd0bdde1238a3de93ac0e444db2976af70811dd27ddda2e1cbd11"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.246513 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr" event={"ID":"3b5de07d-ee5e-492a-9b9d-26de9837ca80","Type":"ContainerStarted","Data":"cb23db601c900262a139721c52e6fec7eb358f4eaf578d98de3cc4eb8dc0924d"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.247166 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.253283 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld" podStartSLOduration=4.208991684 podStartE2EDuration="16.253267456s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:09.817256981 +0000 UTC m=+935.473239455" lastFinishedPulling="2025-11-24 07:03:21.861532753 +0000 UTC m=+947.517515227" observedRunningTime="2025-11-24 07:03:24.248966434 +0000 UTC m=+949.904948908" watchObservedRunningTime="2025-11-24 07:03:24.253267456 +0000 UTC m=+949.909249940" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.255979 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl" event={"ID":"4866479c-d0b9-4366-8ab4-f041573d0d35","Type":"ContainerStarted","Data":"1b53b0db57a69275e8c8e2621c1f80009f81bd29f5e7804cc8cf3eec68b6ee28"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.256348 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.260762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn" event={"ID":"d6a55aa5-3e1c-49e9-bd56-48f8497c1b49","Type":"ContainerStarted","Data":"fed202721e731cc8aafaa8c7349b92645bd3ece1ffae0f7ad1d5430b1cd0571e"} Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.260966 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.272161 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr" podStartSLOduration=4.740627171 podStartE2EDuration="16.272130799s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.331012403 +0000 UTC m=+935.986994877" lastFinishedPulling="2025-11-24 07:03:21.862516031 +0000 UTC m=+947.518498505" observedRunningTime="2025-11-24 07:03:24.267892919 +0000 UTC m=+949.923875393" watchObservedRunningTime="2025-11-24 07:03:24.272130799 +0000 UTC m=+949.928113313" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.326289 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" podStartSLOduration=4.973815118 podStartE2EDuration="16.32626354s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.510761908 +0000 UTC m=+936.166744382" lastFinishedPulling="2025-11-24 07:03:21.86321033 +0000 UTC m=+947.519192804" observedRunningTime="2025-11-24 07:03:24.300358288 +0000 UTC m=+949.956340762" watchObservedRunningTime="2025-11-24 07:03:24.32626354 +0000 UTC m=+949.982246014" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.328137 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt" podStartSLOduration=5.212778787 podStartE2EDuration="16.328130513s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.761980684 +0000 UTC m=+936.417963148" lastFinishedPulling="2025-11-24 07:03:21.8773324 +0000 UTC m=+947.533314874" observedRunningTime="2025-11-24 07:03:24.324976324 +0000 UTC m=+949.980958788" watchObservedRunningTime="2025-11-24 07:03:24.328130513 +0000 UTC m=+949.984112987" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.346536 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2" podStartSLOduration=4.777488744 podStartE2EDuration="16.346515033s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.292403361 +0000 UTC m=+935.948385835" lastFinishedPulling="2025-11-24 07:03:21.86142965 +0000 UTC m=+947.517412124" observedRunningTime="2025-11-24 07:03:24.340723499 +0000 UTC m=+949.996705973" watchObservedRunningTime="2025-11-24 07:03:24.346515033 +0000 UTC m=+950.002497497" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.371759 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8" podStartSLOduration=5.403637735 podStartE2EDuration="16.371737647s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.894787889 +0000 UTC m=+936.550770363" lastFinishedPulling="2025-11-24 07:03:21.862887801 +0000 UTC m=+947.518870275" observedRunningTime="2025-11-24 07:03:24.365187331 +0000 UTC m=+950.021169805" watchObservedRunningTime="2025-11-24 07:03:24.371737647 +0000 UTC m=+950.027720121" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.390595 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd" podStartSLOduration=4.854701769 podStartE2EDuration="16.390560149s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.325688433 +0000 UTC m=+935.981670907" lastFinishedPulling="2025-11-24 07:03:21.861546823 +0000 UTC m=+947.517529287" observedRunningTime="2025-11-24 07:03:24.384400235 +0000 UTC m=+950.040382709" watchObservedRunningTime="2025-11-24 07:03:24.390560149 +0000 UTC m=+950.046542623" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.411679 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt" podStartSLOduration=4.879921161 podStartE2EDuration="16.411645475s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.329715916 +0000 UTC m=+935.985698390" lastFinishedPulling="2025-11-24 07:03:21.86144023 +0000 UTC m=+947.517422704" observedRunningTime="2025-11-24 07:03:24.409113574 +0000 UTC m=+950.065096058" watchObservedRunningTime="2025-11-24 07:03:24.411645475 +0000 UTC m=+950.067627949" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.435502 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf" podStartSLOduration=5.33133646 podStartE2EDuration="16.435468699s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.762323623 +0000 UTC m=+936.418306097" lastFinishedPulling="2025-11-24 07:03:21.866455862 +0000 UTC m=+947.522438336" observedRunningTime="2025-11-24 07:03:24.429948083 +0000 UTC m=+950.085930557" watchObservedRunningTime="2025-11-24 07:03:24.435468699 +0000 UTC m=+950.091451173" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.454722 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9" podStartSLOduration=4.920144919 podStartE2EDuration="16.454696563s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.327037711 +0000 UTC m=+935.983020185" lastFinishedPulling="2025-11-24 07:03:21.861589365 +0000 UTC m=+947.517571829" observedRunningTime="2025-11-24 07:03:24.448840228 +0000 UTC m=+950.104822702" watchObservedRunningTime="2025-11-24 07:03:24.454696563 +0000 UTC m=+950.110679037" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.478315 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw" podStartSLOduration=5.350668167 podStartE2EDuration="16.478246739s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.735934907 +0000 UTC m=+936.391917391" lastFinishedPulling="2025-11-24 07:03:21.863513449 +0000 UTC m=+947.519495963" observedRunningTime="2025-11-24 07:03:24.466287771 +0000 UTC m=+950.122270245" watchObservedRunningTime="2025-11-24 07:03:24.478246739 +0000 UTC m=+950.134229213" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.489489 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx" podStartSLOduration=4.6000945170000005 podStartE2EDuration="16.489472917s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:09.972170143 +0000 UTC m=+935.628152617" lastFinishedPulling="2025-11-24 07:03:21.861548543 +0000 UTC m=+947.517531017" observedRunningTime="2025-11-24 07:03:24.48817051 +0000 UTC m=+950.144152984" watchObservedRunningTime="2025-11-24 07:03:24.489472917 +0000 UTC m=+950.145455391" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.527177 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc" podStartSLOduration=5.198868794 podStartE2EDuration="16.527150533s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.53806368 +0000 UTC m=+936.194046154" lastFinishedPulling="2025-11-24 07:03:21.866345419 +0000 UTC m=+947.522327893" observedRunningTime="2025-11-24 07:03:24.525463725 +0000 UTC m=+950.181446199" watchObservedRunningTime="2025-11-24 07:03:24.527150533 +0000 UTC m=+950.183133017" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.563885 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl" podStartSLOduration=4.735354822 podStartE2EDuration="16.563857961s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.03782608 +0000 UTC m=+935.693808554" lastFinishedPulling="2025-11-24 07:03:21.866329219 +0000 UTC m=+947.522311693" observedRunningTime="2025-11-24 07:03:24.558001495 +0000 UTC m=+950.213983979" watchObservedRunningTime="2025-11-24 07:03:24.563857961 +0000 UTC m=+950.219840435" Nov 24 07:03:24 crc kubenswrapper[4799]: I1124 07:03:24.581642 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn" podStartSLOduration=5.486230161 podStartE2EDuration="16.581622223s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.767441928 +0000 UTC m=+936.423424402" lastFinishedPulling="2025-11-24 07:03:21.86283399 +0000 UTC m=+947.518816464" observedRunningTime="2025-11-24 07:03:24.579135483 +0000 UTC m=+950.235117967" watchObservedRunningTime="2025-11-24 07:03:24.581622223 +0000 UTC m=+950.237604697" Nov 24 07:03:25 crc kubenswrapper[4799]: I1124 07:03:25.317332 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc" Nov 24 07:03:25 crc kubenswrapper[4799]: I1124 07:03:25.317406 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf" Nov 24 07:03:28 crc kubenswrapper[4799]: I1124 07:03:28.564637 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-w7jld" Nov 24 07:03:28 crc kubenswrapper[4799]: I1124 07:03:28.629132 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-rxlcx" Nov 24 07:03:28 crc kubenswrapper[4799]: I1124 07:03:28.631702 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-9prrl" Nov 24 07:03:28 crc kubenswrapper[4799]: I1124 07:03:28.715971 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-9x9d2" Nov 24 07:03:28 crc kubenswrapper[4799]: I1124 07:03:28.827102 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kzmqt" Nov 24 07:03:28 crc kubenswrapper[4799]: I1124 07:03:28.877221 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-nwftw" Nov 24 07:03:28 crc kubenswrapper[4799]: I1124 07:03:28.966232 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-2zmvd" Nov 24 07:03:29 crc kubenswrapper[4799]: I1124 07:03:29.000109 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-mjdkn" Nov 24 07:03:29 crc kubenswrapper[4799]: I1124 07:03:29.001898 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-s7z8n" Nov 24 07:03:29 crc kubenswrapper[4799]: I1124 07:03:29.081634 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-kxlc9" Nov 24 07:03:29 crc kubenswrapper[4799]: I1124 07:03:29.154876 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-2lccf" Nov 24 07:03:29 crc kubenswrapper[4799]: I1124 07:03:29.256582 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-5v5lr" Nov 24 07:03:29 crc kubenswrapper[4799]: I1124 07:03:29.277251 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-k7jkc" Nov 24 07:03:29 crc kubenswrapper[4799]: I1124 07:03:29.556819 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-74qt8" Nov 24 07:03:29 crc kubenswrapper[4799]: I1124 07:03:29.763239 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-fghbt" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.396117 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" event={"ID":"8a1443d1-4bea-481d-ad92-c92e7f5ac7cd","Type":"ContainerStarted","Data":"8c9e3d19f44fe796fba7858b80e519f4013c5576b798b28eae3ab6828cd877a0"} Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.397113 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.400708 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d" event={"ID":"611f4120-8c1b-4745-a5d7-4e1dd52ec567","Type":"ContainerStarted","Data":"f6d501f14d20a93ce3544663722a340d3eb76559e1aa381bd09aed10219144c0"} Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.403560 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" event={"ID":"1d093c2b-22f5-4146-a0e1-a29ab68e1ab3","Type":"ContainerStarted","Data":"3e740ef10c3f46ea061fd7d57667287c619a0b78f53f80cce718bae2229a4b62"} Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.403905 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.407396 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" event={"ID":"b0be97de-cd56-47c0-bd50-868a8fba50d6","Type":"ContainerStarted","Data":"e652e6f338003734b7867a886f0b1d4d98a08d5e37e2803c6f380812d55d692b"} Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.408385 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.412175 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" event={"ID":"66c8f0b0-71e8-4bf6-a28b-109bff76c35e","Type":"ContainerStarted","Data":"ceee3a245ee417f4a56f64ab660c60b8de17f508a6238146b92268ca5c01537e"} Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.412499 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.415445 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" event={"ID":"486bc897-b77f-41d9-8507-a667d595eb61","Type":"ContainerStarted","Data":"44b55ee2e746be5e6df378e4507a6026db2617c768f462a605462d3be098cc7b"} Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.415840 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.417942 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" event={"ID":"fc6a7978-e9a3-4263-89d9-7ed03da8035f","Type":"ContainerStarted","Data":"ed118641ed7ee6ed9655bdfaae3d179d7c91917b5b0811bfa3cf597642c13f21"} Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.418174 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.426273 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" podStartSLOduration=3.772581579 podStartE2EDuration="27.426250352s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.778017467 +0000 UTC m=+936.433999941" lastFinishedPulling="2025-11-24 07:03:34.43168623 +0000 UTC m=+960.087668714" observedRunningTime="2025-11-24 07:03:35.421706024 +0000 UTC m=+961.077688488" watchObservedRunningTime="2025-11-24 07:03:35.426250352 +0000 UTC m=+961.082232826" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.446613 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" podStartSLOduration=4.404134794 podStartE2EDuration="27.446591758s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.799599588 +0000 UTC m=+936.455582062" lastFinishedPulling="2025-11-24 07:03:33.842056542 +0000 UTC m=+959.498039026" observedRunningTime="2025-11-24 07:03:35.441984947 +0000 UTC m=+961.097967421" watchObservedRunningTime="2025-11-24 07:03:35.446591758 +0000 UTC m=+961.102574232" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.463713 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gv46d" podStartSLOduration=2.897223854 podStartE2EDuration="26.463688391s" podCreationTimestamp="2025-11-24 07:03:09 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.936346145 +0000 UTC m=+936.592328619" lastFinishedPulling="2025-11-24 07:03:34.502810682 +0000 UTC m=+960.158793156" observedRunningTime="2025-11-24 07:03:35.457425054 +0000 UTC m=+961.113407528" watchObservedRunningTime="2025-11-24 07:03:35.463688391 +0000 UTC m=+961.119670865" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.488590 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" podStartSLOduration=4.036411271 podStartE2EDuration="27.488566265s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.932548927 +0000 UTC m=+936.588531401" lastFinishedPulling="2025-11-24 07:03:34.384703911 +0000 UTC m=+960.040686395" observedRunningTime="2025-11-24 07:03:35.482897615 +0000 UTC m=+961.138880089" watchObservedRunningTime="2025-11-24 07:03:35.488566265 +0000 UTC m=+961.144548739" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.524675 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" podStartSLOduration=4.479927618 podStartE2EDuration="27.524649696s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.798177507 +0000 UTC m=+936.454159991" lastFinishedPulling="2025-11-24 07:03:33.842899595 +0000 UTC m=+959.498882069" observedRunningTime="2025-11-24 07:03:35.523427301 +0000 UTC m=+961.179409795" watchObservedRunningTime="2025-11-24 07:03:35.524649696 +0000 UTC m=+961.180632170" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.552119 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" podStartSLOduration=4.05404936 podStartE2EDuration="27.552096272s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.933290678 +0000 UTC m=+936.589273152" lastFinishedPulling="2025-11-24 07:03:34.43133759 +0000 UTC m=+960.087320064" observedRunningTime="2025-11-24 07:03:35.546454852 +0000 UTC m=+961.202437326" watchObservedRunningTime="2025-11-24 07:03:35.552096272 +0000 UTC m=+961.208078746" Nov 24 07:03:35 crc kubenswrapper[4799]: I1124 07:03:35.564834 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" podStartSLOduration=4.012408252 podStartE2EDuration="27.564812592s" podCreationTimestamp="2025-11-24 07:03:08 +0000 UTC" firstStartedPulling="2025-11-24 07:03:10.932834675 +0000 UTC m=+936.588817149" lastFinishedPulling="2025-11-24 07:03:34.485239005 +0000 UTC m=+960.141221489" observedRunningTime="2025-11-24 07:03:35.562015592 +0000 UTC m=+961.217998056" watchObservedRunningTime="2025-11-24 07:03:35.564812592 +0000 UTC m=+961.220795066" Nov 24 07:03:39 crc kubenswrapper[4799]: I1124 07:03:39.196779 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-xgq2g" Nov 24 07:03:39 crc kubenswrapper[4799]: I1124 07:03:39.315484 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2xzbc" Nov 24 07:03:39 crc kubenswrapper[4799]: I1124 07:03:39.611655 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-scbrc" Nov 24 07:03:39 crc kubenswrapper[4799]: I1124 07:03:39.641984 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-d4b8t" Nov 24 07:03:39 crc kubenswrapper[4799]: I1124 07:03:39.719147 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-8464cf66df-9rph7" Nov 24 07:03:40 crc kubenswrapper[4799]: I1124 07:03:40.188744 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5" Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.904060 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-b8lv4"] Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.914496 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.917822 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.918086 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.918287 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.918680 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-bgtz9" Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.920059 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-b8lv4"] Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.972572 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg8cz\" (UniqueName: \"kubernetes.io/projected/bfb1e96e-5d13-4725-8654-b9235a261a22-kube-api-access-vg8cz\") pod \"dnsmasq-dns-7bdd77c89-b8lv4\" (UID: \"bfb1e96e-5d13-4725-8654-b9235a261a22\") " pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.972650 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb1e96e-5d13-4725-8654-b9235a261a22-config\") pod \"dnsmasq-dns-7bdd77c89-b8lv4\" (UID: \"bfb1e96e-5d13-4725-8654-b9235a261a22\") " pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.988739 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6584b49599-r5d9g"] Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.990273 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:03:56 crc kubenswrapper[4799]: I1124 07:03:56.993715 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.005996 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-r5d9g"] Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.073908 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9rdr\" (UniqueName: \"kubernetes.io/projected/7b92b41b-1fff-408e-96a9-b138e8ffa320-kube-api-access-v9rdr\") pod \"dnsmasq-dns-6584b49599-r5d9g\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.074136 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-config\") pod \"dnsmasq-dns-6584b49599-r5d9g\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.074244 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg8cz\" (UniqueName: \"kubernetes.io/projected/bfb1e96e-5d13-4725-8654-b9235a261a22-kube-api-access-vg8cz\") pod \"dnsmasq-dns-7bdd77c89-b8lv4\" (UID: \"bfb1e96e-5d13-4725-8654-b9235a261a22\") " pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.074338 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-dns-svc\") pod \"dnsmasq-dns-6584b49599-r5d9g\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.074416 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb1e96e-5d13-4725-8654-b9235a261a22-config\") pod \"dnsmasq-dns-7bdd77c89-b8lv4\" (UID: \"bfb1e96e-5d13-4725-8654-b9235a261a22\") " pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.075416 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb1e96e-5d13-4725-8654-b9235a261a22-config\") pod \"dnsmasq-dns-7bdd77c89-b8lv4\" (UID: \"bfb1e96e-5d13-4725-8654-b9235a261a22\") " pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.106991 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg8cz\" (UniqueName: \"kubernetes.io/projected/bfb1e96e-5d13-4725-8654-b9235a261a22-kube-api-access-vg8cz\") pod \"dnsmasq-dns-7bdd77c89-b8lv4\" (UID: \"bfb1e96e-5d13-4725-8654-b9235a261a22\") " pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.176402 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9rdr\" (UniqueName: \"kubernetes.io/projected/7b92b41b-1fff-408e-96a9-b138e8ffa320-kube-api-access-v9rdr\") pod \"dnsmasq-dns-6584b49599-r5d9g\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.176508 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-config\") pod \"dnsmasq-dns-6584b49599-r5d9g\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.176567 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-dns-svc\") pod \"dnsmasq-dns-6584b49599-r5d9g\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.177833 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-dns-svc\") pod \"dnsmasq-dns-6584b49599-r5d9g\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.177870 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-config\") pod \"dnsmasq-dns-6584b49599-r5d9g\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.193488 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9rdr\" (UniqueName: \"kubernetes.io/projected/7b92b41b-1fff-408e-96a9-b138e8ffa320-kube-api-access-v9rdr\") pod \"dnsmasq-dns-6584b49599-r5d9g\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.242439 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.306963 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.715156 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-b8lv4"] Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.723427 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:03:57 crc kubenswrapper[4799]: I1124 07:03:57.796931 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-r5d9g"] Nov 24 07:03:57 crc kubenswrapper[4799]: W1124 07:03:57.804804 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b92b41b_1fff_408e_96a9_b138e8ffa320.slice/crio-cc37be54ebac6831c011ff270ff1bfb318ee84a8220feaf4cbb46fba09464ea0 WatchSource:0}: Error finding container cc37be54ebac6831c011ff270ff1bfb318ee84a8220feaf4cbb46fba09464ea0: Status 404 returned error can't find the container with id cc37be54ebac6831c011ff270ff1bfb318ee84a8220feaf4cbb46fba09464ea0 Nov 24 07:03:58 crc kubenswrapper[4799]: I1124 07:03:58.635381 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" event={"ID":"bfb1e96e-5d13-4725-8654-b9235a261a22","Type":"ContainerStarted","Data":"e7a84c7abde508e578c9076291cfc6f65645a0300331c62f3c8eb403f524c7e9"} Nov 24 07:03:58 crc kubenswrapper[4799]: I1124 07:03:58.638683 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-r5d9g" event={"ID":"7b92b41b-1fff-408e-96a9-b138e8ffa320","Type":"ContainerStarted","Data":"cc37be54ebac6831c011ff270ff1bfb318ee84a8220feaf4cbb46fba09464ea0"} Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.085548 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-r5d9g"] Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.110277 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-7np9d"] Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.116260 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.131926 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-7np9d"] Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.222823 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm97x\" (UniqueName: \"kubernetes.io/projected/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-kube-api-access-nm97x\") pod \"dnsmasq-dns-7c6d9948dc-7np9d\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.223010 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-config\") pod \"dnsmasq-dns-7c6d9948dc-7np9d\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.223073 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-7np9d\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.328045 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-7np9d\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.328100 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm97x\" (UniqueName: \"kubernetes.io/projected/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-kube-api-access-nm97x\") pod \"dnsmasq-dns-7c6d9948dc-7np9d\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.328186 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-config\") pod \"dnsmasq-dns-7c6d9948dc-7np9d\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.332124 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-config\") pod \"dnsmasq-dns-7c6d9948dc-7np9d\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.337214 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-7np9d\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.379703 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm97x\" (UniqueName: \"kubernetes.io/projected/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-kube-api-access-nm97x\") pod \"dnsmasq-dns-7c6d9948dc-7np9d\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.431825 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-b8lv4"] Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.437437 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.447367 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-h94gz"] Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.448783 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.492523 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-h94gz"] Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.532444 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-dns-svc\") pod \"dnsmasq-dns-6486446b9f-h94gz\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.532493 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-config\") pod \"dnsmasq-dns-6486446b9f-h94gz\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.532558 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzb2m\" (UniqueName: \"kubernetes.io/projected/0285d647-978a-4e9b-8ba2-b369a038db04-kube-api-access-vzb2m\") pod \"dnsmasq-dns-6486446b9f-h94gz\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.633881 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-dns-svc\") pod \"dnsmasq-dns-6486446b9f-h94gz\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.634335 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-config\") pod \"dnsmasq-dns-6486446b9f-h94gz\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.634398 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzb2m\" (UniqueName: \"kubernetes.io/projected/0285d647-978a-4e9b-8ba2-b369a038db04-kube-api-access-vzb2m\") pod \"dnsmasq-dns-6486446b9f-h94gz\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.635600 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-dns-svc\") pod \"dnsmasq-dns-6486446b9f-h94gz\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.640973 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-config\") pod \"dnsmasq-dns-6486446b9f-h94gz\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.679158 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzb2m\" (UniqueName: \"kubernetes.io/projected/0285d647-978a-4e9b-8ba2-b369a038db04-kube-api-access-vzb2m\") pod \"dnsmasq-dns-6486446b9f-h94gz\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:03:59 crc kubenswrapper[4799]: I1124 07:03:59.872251 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.158880 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-7np9d"] Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.296030 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.297441 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.301063 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-khxp8" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.301436 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.301672 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.302047 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.302420 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.302912 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.308710 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.314599 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.348276 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.348400 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/96acf88b-3fe8-46ac-8393-37f720632dd6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.348447 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/96acf88b-3fe8-46ac-8393-37f720632dd6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.348524 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.348588 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.348663 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh8mr\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-kube-api-access-wh8mr\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.348694 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.348860 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.348904 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.348929 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.348947 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.350885 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-h94gz"] Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.450221 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.450303 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.450331 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.450357 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.450396 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/96acf88b-3fe8-46ac-8393-37f720632dd6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.450424 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/96acf88b-3fe8-46ac-8393-37f720632dd6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.450446 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.450468 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.450483 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh8mr\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-kube-api-access-wh8mr\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.450519 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.450577 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.451151 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.451435 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.452937 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.453286 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.458517 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.453553 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.462222 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.462788 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/96acf88b-3fe8-46ac-8393-37f720632dd6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.466310 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/96acf88b-3fe8-46ac-8393-37f720632dd6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.467895 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.477571 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh8mr\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-kube-api-access-wh8mr\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.495920 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.629922 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.658553 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.665471 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.671197 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.672308 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.672597 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.673456 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.673779 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.673802 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.687808 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5qjfb" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.709544 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.733091 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" event={"ID":"ae46b5b6-4f76-4730-923e-7c44c7eaafbe","Type":"ContainerStarted","Data":"2c7e172f4f7929289607d7e0441e5bf7d8189638aaea82ce127cfa8a5c30a430"} Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.757702 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.757769 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.757826 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.757875 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.757926 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.757952 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/366b7548-2fca-4623-a45c-c1e0367ce93a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.757976 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwkm6\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-kube-api-access-dwkm6\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.758008 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.758063 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.758112 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.758145 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/366b7548-2fca-4623-a45c-c1e0367ce93a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.768040 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" event={"ID":"0285d647-978a-4e9b-8ba2-b369a038db04","Type":"ContainerStarted","Data":"c4563b2aaf818bc7d31b548f44b49b4f543d70d369fe9a30546f0694edca1853"} Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.859644 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.859697 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/366b7548-2fca-4623-a45c-c1e0367ce93a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.861138 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwkm6\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-kube-api-access-dwkm6\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.861181 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.861212 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.861267 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.861301 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/366b7548-2fca-4623-a45c-c1e0367ce93a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.861352 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.862059 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.862551 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.863151 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.867942 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/366b7548-2fca-4623-a45c-c1e0367ce93a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.868483 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.868570 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.869216 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.869272 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.869655 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.869680 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.870303 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.871667 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.876397 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/366b7548-2fca-4623-a45c-c1e0367ce93a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.882019 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwkm6\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-kube-api-access-dwkm6\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:00 crc kubenswrapper[4799]: I1124 07:04:00.910021 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.045534 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.277100 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:04:01 crc kubenswrapper[4799]: W1124 07:04:01.296467 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96acf88b_3fe8_46ac_8393_37f720632dd6.slice/crio-be3cdda108578f8a4d7197ec7c3ca0efe2b97926c61c400bde868bb644f21f36 WatchSource:0}: Error finding container be3cdda108578f8a4d7197ec7c3ca0efe2b97926c61c400bde868bb644f21f36: Status 404 returned error can't find the container with id be3cdda108578f8a4d7197ec7c3ca0efe2b97926c61c400bde868bb644f21f36 Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.614487 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.777301 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"366b7548-2fca-4623-a45c-c1e0367ce93a","Type":"ContainerStarted","Data":"602ba2b5867b8495e6b82dde9d4cfe4dedbf52572222ca4badcb957c182c94c2"} Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.778833 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"96acf88b-3fe8-46ac-8393-37f720632dd6","Type":"ContainerStarted","Data":"be3cdda108578f8a4d7197ec7c3ca0efe2b97926c61c400bde868bb644f21f36"} Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.850973 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.855302 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.860345 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.862757 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.863417 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.864141 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.864219 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-hptzz" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.872314 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.888239 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kolla-config\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.888334 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.888390 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.888444 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-operator-scripts\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.888503 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkvnz\" (UniqueName: \"kubernetes.io/projected/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kube-api-access-dkvnz\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.888540 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-generated\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.888567 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.888593 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-default\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.990226 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkvnz\" (UniqueName: \"kubernetes.io/projected/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kube-api-access-dkvnz\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.990289 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-generated\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.990316 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.990338 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-default\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.990388 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kolla-config\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.990415 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.990464 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.990504 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-operator-scripts\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.992471 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kolla-config\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.992938 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.992946 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-generated\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.993106 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-operator-scripts\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:01 crc kubenswrapper[4799]: I1124 07:04:01.996450 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-default\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:02 crc kubenswrapper[4799]: I1124 07:04:02.014677 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:02 crc kubenswrapper[4799]: I1124 07:04:02.014761 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:02 crc kubenswrapper[4799]: I1124 07:04:02.021284 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkvnz\" (UniqueName: \"kubernetes.io/projected/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kube-api-access-dkvnz\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:02 crc kubenswrapper[4799]: I1124 07:04:02.033641 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " pod="openstack/openstack-galera-0" Nov 24 07:04:02 crc kubenswrapper[4799]: I1124 07:04:02.195880 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.288040 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.290531 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.294882 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.295247 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-vcwvn" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.295503 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.299407 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.301079 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.439390 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.439464 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd7rz\" (UniqueName: \"kubernetes.io/projected/844d53db-a457-4153-bdb4-d557ba34c099-kube-api-access-gd7rz\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.439489 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.439519 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.439564 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.439602 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/844d53db-a457-4153-bdb4-d557ba34c099-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.439638 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.439663 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.458401 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.459879 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.464427 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.464813 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.464989 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-25ltl" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.474727 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541540 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541610 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/844d53db-a457-4153-bdb4-d557ba34c099-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541649 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-kolla-config\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541673 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541691 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541731 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-config-data\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541751 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541772 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541799 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd7rz\" (UniqueName: \"kubernetes.io/projected/844d53db-a457-4153-bdb4-d557ba34c099-kube-api-access-gd7rz\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541817 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541869 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541889 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl9kp\" (UniqueName: \"kubernetes.io/projected/b0361baf-d0c3-430f-87e3-c011f869a2b3-kube-api-access-jl9kp\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.541921 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.542471 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.543148 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.543691 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.544022 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.544696 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/844d53db-a457-4153-bdb4-d557ba34c099-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.551586 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.552469 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.580327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.581241 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd7rz\" (UniqueName: \"kubernetes.io/projected/844d53db-a457-4153-bdb4-d557ba34c099-kube-api-access-gd7rz\") pod \"openstack-cell1-galera-0\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.610537 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.643675 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl9kp\" (UniqueName: \"kubernetes.io/projected/b0361baf-d0c3-430f-87e3-c011f869a2b3-kube-api-access-jl9kp\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.643797 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.643965 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-kolla-config\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.644535 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-config-data\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.644594 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.648747 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.648899 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-kolla-config\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.649488 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-config-data\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.651950 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.669606 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl9kp\" (UniqueName: \"kubernetes.io/projected/b0361baf-d0c3-430f-87e3-c011f869a2b3-kube-api-access-jl9kp\") pod \"memcached-0\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " pod="openstack/memcached-0" Nov 24 07:04:03 crc kubenswrapper[4799]: I1124 07:04:03.801820 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 07:04:05 crc kubenswrapper[4799]: I1124 07:04:05.550944 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:04:05 crc kubenswrapper[4799]: I1124 07:04:05.552135 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:04:05 crc kubenswrapper[4799]: I1124 07:04:05.555370 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-wfgc7" Nov 24 07:04:05 crc kubenswrapper[4799]: I1124 07:04:05.566511 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:04:05 crc kubenswrapper[4799]: I1124 07:04:05.587289 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbwwq\" (UniqueName: \"kubernetes.io/projected/04342f4b-48ae-4c45-ad6b-947b30b9de75-kube-api-access-bbwwq\") pod \"kube-state-metrics-0\" (UID: \"04342f4b-48ae-4c45-ad6b-947b30b9de75\") " pod="openstack/kube-state-metrics-0" Nov 24 07:04:05 crc kubenswrapper[4799]: I1124 07:04:05.692903 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbwwq\" (UniqueName: \"kubernetes.io/projected/04342f4b-48ae-4c45-ad6b-947b30b9de75-kube-api-access-bbwwq\") pod \"kube-state-metrics-0\" (UID: \"04342f4b-48ae-4c45-ad6b-947b30b9de75\") " pod="openstack/kube-state-metrics-0" Nov 24 07:04:05 crc kubenswrapper[4799]: I1124 07:04:05.745271 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbwwq\" (UniqueName: \"kubernetes.io/projected/04342f4b-48ae-4c45-ad6b-947b30b9de75-kube-api-access-bbwwq\") pod \"kube-state-metrics-0\" (UID: \"04342f4b-48ae-4c45-ad6b-947b30b9de75\") " pod="openstack/kube-state-metrics-0" Nov 24 07:04:05 crc kubenswrapper[4799]: I1124 07:04:05.870163 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.692634 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.700326 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.707510 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.719815 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.720135 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.720252 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.720359 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.720498 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-4cw24" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.846411 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2dz2n"] Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.847933 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.852536 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-lk9bx" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.852774 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.853840 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.864181 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2dz2n"] Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.872428 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drg4w\" (UniqueName: \"kubernetes.io/projected/bbfb60f9-6e71-4b22-9d74-cab607a00c20-kube-api-access-drg4w\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.872532 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.872611 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.872649 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-config\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.872872 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.872903 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.872923 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.872945 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.882044 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-kgrrs"] Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.883990 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.902874 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-kgrrs"] Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.974279 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jrkq\" (UniqueName: \"kubernetes.io/projected/dd32bea6-7ad9-4182-bc59-5644efe24ce1-kube-api-access-6jrkq\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.974349 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-combined-ca-bundle\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.974599 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd32bea6-7ad9-4182-bc59-5644efe24ce1-scripts\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.974758 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drg4w\" (UniqueName: \"kubernetes.io/projected/bbfb60f9-6e71-4b22-9d74-cab607a00c20-kube-api-access-drg4w\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.974953 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-ovn-controller-tls-certs\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.975000 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.975023 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run-ovn\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.975371 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.975437 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-config\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.975479 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.975541 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-log-ovn\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.975606 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.975706 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.976160 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.976276 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.976350 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.976587 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-config\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.977419 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.978008 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.985027 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.985295 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.985529 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:09 crc kubenswrapper[4799]: I1124 07:04:09.998259 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drg4w\" (UniqueName: \"kubernetes.io/projected/bbfb60f9-6e71-4b22-9d74-cab607a00c20-kube-api-access-drg4w\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.021916 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.065504 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.084712 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-log-ovn\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.084777 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4nd7\" (UniqueName: \"kubernetes.io/projected/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-kube-api-access-r4nd7\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.084804 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.085043 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jrkq\" (UniqueName: \"kubernetes.io/projected/dd32bea6-7ad9-4182-bc59-5644efe24ce1-kube-api-access-6jrkq\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.085089 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-log\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.085116 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-run\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.085147 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-combined-ca-bundle\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.085179 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-lib\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.085206 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-etc-ovs\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.085240 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd32bea6-7ad9-4182-bc59-5644efe24ce1-scripts\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.085276 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-ovn-controller-tls-certs\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.085297 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run-ovn\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.085332 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-scripts\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.086445 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-log-ovn\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.091022 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run-ovn\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.092437 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd32bea6-7ad9-4182-bc59-5644efe24ce1-scripts\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.093409 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-ovn-controller-tls-certs\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.093657 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-combined-ca-bundle\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.095584 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.105242 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jrkq\" (UniqueName: \"kubernetes.io/projected/dd32bea6-7ad9-4182-bc59-5644efe24ce1-kube-api-access-6jrkq\") pod \"ovn-controller-2dz2n\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.186960 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-log\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.187015 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-run\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.187059 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-lib\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.187088 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-etc-ovs\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.187144 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-scripts\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.187144 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-run\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.187182 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4nd7\" (UniqueName: \"kubernetes.io/projected/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-kube-api-access-r4nd7\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.187231 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-log\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.187258 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-lib\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.187360 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-etc-ovs\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.189715 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-scripts\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.206713 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4nd7\" (UniqueName: \"kubernetes.io/projected/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-kube-api-access-r4nd7\") pod \"ovn-controller-ovs-kgrrs\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.207753 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:10 crc kubenswrapper[4799]: I1124 07:04:10.215130 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.619655 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.621783 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.630698 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.631070 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.631734 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vgjxs" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.632025 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.640818 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.737145 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.737225 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.737295 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.737359 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.737452 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79m49\" (UniqueName: \"kubernetes.io/projected/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-kube-api-access-79m49\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.737486 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.737555 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-config\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.737824 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.840329 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.840394 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.840449 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.840474 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79m49\" (UniqueName: \"kubernetes.io/projected/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-kube-api-access-79m49\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.840524 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-config\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.840590 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.840663 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.840690 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.840826 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.841786 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.842572 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.847870 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-config\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.849176 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.853510 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.866657 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.871293 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.876869 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79m49\" (UniqueName: \"kubernetes.io/projected/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-kube-api-access-79m49\") pod \"ovsdbserver-sb-0\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:12 crc kubenswrapper[4799]: I1124 07:04:12.956288 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:16 crc kubenswrapper[4799]: E1124 07:04:16.474246 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 24 07:04:16 crc kubenswrapper[4799]: E1124 07:04:16.475371 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vg8cz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bdd77c89-b8lv4_openstack(bfb1e96e-5d13-4725-8654-b9235a261a22): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:04:16 crc kubenswrapper[4799]: E1124 07:04:16.476580 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" podUID="bfb1e96e-5d13-4725-8654-b9235a261a22" Nov 24 07:04:16 crc kubenswrapper[4799]: E1124 07:04:16.497232 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 24 07:04:16 crc kubenswrapper[4799]: E1124 07:04:16.497607 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v9rdr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6584b49599-r5d9g_openstack(7b92b41b-1fff-408e-96a9-b138e8ffa320): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:04:16 crc kubenswrapper[4799]: E1124 07:04:16.499109 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6584b49599-r5d9g" podUID="7b92b41b-1fff-408e-96a9-b138e8ffa320" Nov 24 07:04:17 crc kubenswrapper[4799]: I1124 07:04:17.884860 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:04:17 crc kubenswrapper[4799]: I1124 07:04:17.907251 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" Nov 24 07:04:17 crc kubenswrapper[4799]: I1124 07:04:17.950922 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" event={"ID":"bfb1e96e-5d13-4725-8654-b9235a261a22","Type":"ContainerDied","Data":"e7a84c7abde508e578c9076291cfc6f65645a0300331c62f3c8eb403f524c7e9"} Nov 24 07:04:17 crc kubenswrapper[4799]: I1124 07:04:17.951054 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-b8lv4" Nov 24 07:04:17 crc kubenswrapper[4799]: I1124 07:04:17.973258 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-r5d9g" event={"ID":"7b92b41b-1fff-408e-96a9-b138e8ffa320","Type":"ContainerDied","Data":"cc37be54ebac6831c011ff270ff1bfb318ee84a8220feaf4cbb46fba09464ea0"} Nov 24 07:04:17 crc kubenswrapper[4799]: I1124 07:04:17.973374 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-r5d9g" Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.078034 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.086242 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9rdr\" (UniqueName: \"kubernetes.io/projected/7b92b41b-1fff-408e-96a9-b138e8ffa320-kube-api-access-v9rdr\") pod \"7b92b41b-1fff-408e-96a9-b138e8ffa320\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.086331 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-config\") pod \"7b92b41b-1fff-408e-96a9-b138e8ffa320\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.086350 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg8cz\" (UniqueName: \"kubernetes.io/projected/bfb1e96e-5d13-4725-8654-b9235a261a22-kube-api-access-vg8cz\") pod \"bfb1e96e-5d13-4725-8654-b9235a261a22\" (UID: \"bfb1e96e-5d13-4725-8654-b9235a261a22\") " Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.086393 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-dns-svc\") pod \"7b92b41b-1fff-408e-96a9-b138e8ffa320\" (UID: \"7b92b41b-1fff-408e-96a9-b138e8ffa320\") " Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.086470 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb1e96e-5d13-4725-8654-b9235a261a22-config\") pod \"bfb1e96e-5d13-4725-8654-b9235a261a22\" (UID: \"bfb1e96e-5d13-4725-8654-b9235a261a22\") " Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.087304 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-config" (OuterVolumeSpecName: "config") pod "7b92b41b-1fff-408e-96a9-b138e8ffa320" (UID: "7b92b41b-1fff-408e-96a9-b138e8ffa320"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.087874 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7b92b41b-1fff-408e-96a9-b138e8ffa320" (UID: "7b92b41b-1fff-408e-96a9-b138e8ffa320"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.087963 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb1e96e-5d13-4725-8654-b9235a261a22-config" (OuterVolumeSpecName: "config") pod "bfb1e96e-5d13-4725-8654-b9235a261a22" (UID: "bfb1e96e-5d13-4725-8654-b9235a261a22"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.091276 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfb1e96e-5d13-4725-8654-b9235a261a22-kube-api-access-vg8cz" (OuterVolumeSpecName: "kube-api-access-vg8cz") pod "bfb1e96e-5d13-4725-8654-b9235a261a22" (UID: "bfb1e96e-5d13-4725-8654-b9235a261a22"). InnerVolumeSpecName "kube-api-access-vg8cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.092560 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b92b41b-1fff-408e-96a9-b138e8ffa320-kube-api-access-v9rdr" (OuterVolumeSpecName: "kube-api-access-v9rdr") pod "7b92b41b-1fff-408e-96a9-b138e8ffa320" (UID: "7b92b41b-1fff-408e-96a9-b138e8ffa320"). InnerVolumeSpecName "kube-api-access-v9rdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.124991 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.188960 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9rdr\" (UniqueName: \"kubernetes.io/projected/7b92b41b-1fff-408e-96a9-b138e8ffa320-kube-api-access-v9rdr\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.189003 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.189015 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg8cz\" (UniqueName: \"kubernetes.io/projected/bfb1e96e-5d13-4725-8654-b9235a261a22-kube-api-access-vg8cz\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.189023 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b92b41b-1fff-408e-96a9-b138e8ffa320-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.189032 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb1e96e-5d13-4725-8654-b9235a261a22-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.311579 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-b8lv4"] Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.314653 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-b8lv4"] Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.353044 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.361582 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-r5d9g"] Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.367732 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-r5d9g"] Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.497109 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.509363 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2dz2n"] Nov 24 07:04:18 crc kubenswrapper[4799]: W1124 07:04:18.525036 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod997ffb9e_cc6a_4fca_91d9_eb17d711b135.slice/crio-da9ce81af35a9eaee55e0fbdda1e265b5056838f3b48fe78debfd2804a6c6596 WatchSource:0}: Error finding container da9ce81af35a9eaee55e0fbdda1e265b5056838f3b48fe78debfd2804a6c6596: Status 404 returned error can't find the container with id da9ce81af35a9eaee55e0fbdda1e265b5056838f3b48fe78debfd2804a6c6596 Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.535268 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 07:04:18 crc kubenswrapper[4799]: W1124 07:04:18.535939 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd32bea6_7ad9_4182_bc59_5644efe24ce1.slice/crio-ca0366e7a6e2436619aa02f7351d51fd3452b3bdb2ad73621e1c3046ac9d75e3 WatchSource:0}: Error finding container ca0366e7a6e2436619aa02f7351d51fd3452b3bdb2ad73621e1c3046ac9d75e3: Status 404 returned error can't find the container with id ca0366e7a6e2436619aa02f7351d51fd3452b3bdb2ad73621e1c3046ac9d75e3 Nov 24 07:04:18 crc kubenswrapper[4799]: W1124 07:04:18.548302 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbfb60f9_6e71_4b22_9d74_cab607a00c20.slice/crio-291ecd9e8f24fbcb763f8e7895bcbae09f1ba8eaa22a90e8c62932f922733e6e WatchSource:0}: Error finding container 291ecd9e8f24fbcb763f8e7895bcbae09f1ba8eaa22a90e8c62932f922733e6e: Status 404 returned error can't find the container with id 291ecd9e8f24fbcb763f8e7895bcbae09f1ba8eaa22a90e8c62932f922733e6e Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.607708 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.985668 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n" event={"ID":"dd32bea6-7ad9-4182-bc59-5644efe24ce1","Type":"ContainerStarted","Data":"ca0366e7a6e2436619aa02f7351d51fd3452b3bdb2ad73621e1c3046ac9d75e3"} Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.987693 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"997ffb9e-cc6a-4fca-91d9-eb17d711b135","Type":"ContainerStarted","Data":"da9ce81af35a9eaee55e0fbdda1e265b5056838f3b48fe78debfd2804a6c6596"} Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.988924 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bbfb60f9-6e71-4b22-9d74-cab607a00c20","Type":"ContainerStarted","Data":"291ecd9e8f24fbcb763f8e7895bcbae09f1ba8eaa22a90e8c62932f922733e6e"} Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.990274 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b0361baf-d0c3-430f-87e3-c011f869a2b3","Type":"ContainerStarted","Data":"0a0a4d0f870ff284bd904a9177b1cc9725b1d472f9f64e6febe1d4f9ae5730c3"} Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.991630 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace","Type":"ContainerStarted","Data":"eb13270ff2460e3f3e8b1c5606ce701360ed59511f174e9b8db1cdf0a8e86f75"} Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.992734 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"844d53db-a457-4153-bdb4-d557ba34c099","Type":"ContainerStarted","Data":"46cd63af2e8051dfb8835b25509abf388109dc011bc4a05c232bafe7a0f0c1e5"} Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.995670 4799 generic.go:334] "Generic (PLEG): container finished" podID="ae46b5b6-4f76-4730-923e-7c44c7eaafbe" containerID="a48042e0db19d83f9c784d1d6228c57c41f55f99545e98f19aef096cc18b2df7" exitCode=0 Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.995795 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" event={"ID":"ae46b5b6-4f76-4730-923e-7c44c7eaafbe","Type":"ContainerDied","Data":"a48042e0db19d83f9c784d1d6228c57c41f55f99545e98f19aef096cc18b2df7"} Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.997929 4799 generic.go:334] "Generic (PLEG): container finished" podID="0285d647-978a-4e9b-8ba2-b369a038db04" containerID="5044f3331cc3cd478cb2dce11a0fec48a73f5eeccf2e4bbc37bf383f6c48534b" exitCode=0 Nov 24 07:04:18 crc kubenswrapper[4799]: I1124 07:04:18.998000 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" event={"ID":"0285d647-978a-4e9b-8ba2-b369a038db04","Type":"ContainerDied","Data":"5044f3331cc3cd478cb2dce11a0fec48a73f5eeccf2e4bbc37bf383f6c48534b"} Nov 24 07:04:19 crc kubenswrapper[4799]: I1124 07:04:19.001270 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"96acf88b-3fe8-46ac-8393-37f720632dd6","Type":"ContainerStarted","Data":"0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29"} Nov 24 07:04:19 crc kubenswrapper[4799]: I1124 07:04:19.003136 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"04342f4b-48ae-4c45-ad6b-947b30b9de75","Type":"ContainerStarted","Data":"80ffe4ba2b05d66cbac89e0520e04e2880c6f22f2ff8744e1caeb369043cff2f"} Nov 24 07:04:19 crc kubenswrapper[4799]: I1124 07:04:19.582125 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-kgrrs"] Nov 24 07:04:19 crc kubenswrapper[4799]: I1124 07:04:19.641785 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b92b41b-1fff-408e-96a9-b138e8ffa320" path="/var/lib/kubelet/pods/7b92b41b-1fff-408e-96a9-b138e8ffa320/volumes" Nov 24 07:04:19 crc kubenswrapper[4799]: I1124 07:04:19.642244 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfb1e96e-5d13-4725-8654-b9235a261a22" path="/var/lib/kubelet/pods/bfb1e96e-5d13-4725-8654-b9235a261a22/volumes" Nov 24 07:04:19 crc kubenswrapper[4799]: W1124 07:04:19.816341 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7742c81_45e0_4ef5_aceb_7a631bf9cc37.slice/crio-d5dbdbfdad4349f320febaff6ac846a95686336f646a6f63854dac9181e3cd33 WatchSource:0}: Error finding container d5dbdbfdad4349f320febaff6ac846a95686336f646a6f63854dac9181e3cd33: Status 404 returned error can't find the container with id d5dbdbfdad4349f320febaff6ac846a95686336f646a6f63854dac9181e3cd33 Nov 24 07:04:20 crc kubenswrapper[4799]: I1124 07:04:20.012430 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kgrrs" event={"ID":"a7742c81-45e0-4ef5-aceb-7a631bf9cc37","Type":"ContainerStarted","Data":"d5dbdbfdad4349f320febaff6ac846a95686336f646a6f63854dac9181e3cd33"} Nov 24 07:04:20 crc kubenswrapper[4799]: I1124 07:04:20.015060 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" event={"ID":"0285d647-978a-4e9b-8ba2-b369a038db04","Type":"ContainerStarted","Data":"0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338"} Nov 24 07:04:20 crc kubenswrapper[4799]: I1124 07:04:20.015207 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:04:20 crc kubenswrapper[4799]: I1124 07:04:20.018633 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"366b7548-2fca-4623-a45c-c1e0367ce93a","Type":"ContainerStarted","Data":"c6e22b9dd219757023e664a05027fc4af7ee03f3c7e62fe2a381ed00aef63570"} Nov 24 07:04:20 crc kubenswrapper[4799]: I1124 07:04:20.039623 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" podStartSLOduration=3.670441186 podStartE2EDuration="21.039594875s" podCreationTimestamp="2025-11-24 07:03:59 +0000 UTC" firstStartedPulling="2025-11-24 07:04:00.374469732 +0000 UTC m=+986.030452206" lastFinishedPulling="2025-11-24 07:04:17.743623421 +0000 UTC m=+1003.399605895" observedRunningTime="2025-11-24 07:04:20.03308192 +0000 UTC m=+1005.689064404" watchObservedRunningTime="2025-11-24 07:04:20.039594875 +0000 UTC m=+1005.695577349" Nov 24 07:04:24 crc kubenswrapper[4799]: I1124 07:04:24.874107 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:04:24 crc kubenswrapper[4799]: I1124 07:04:24.953944 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-7np9d"] Nov 24 07:04:25 crc kubenswrapper[4799]: E1124 07:04:25.758034 4799 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 24 07:04:25 crc kubenswrapper[4799]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/ae46b5b6-4f76-4730-923e-7c44c7eaafbe/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 24 07:04:25 crc kubenswrapper[4799]: > podSandboxID="2c7e172f4f7929289607d7e0441e5bf7d8189638aaea82ce127cfa8a5c30a430" Nov 24 07:04:25 crc kubenswrapper[4799]: E1124 07:04:25.758674 4799 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 24 07:04:25 crc kubenswrapper[4799]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nm97x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7c6d9948dc-7np9d_openstack(ae46b5b6-4f76-4730-923e-7c44c7eaafbe): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/ae46b5b6-4f76-4730-923e-7c44c7eaafbe/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 24 07:04:25 crc kubenswrapper[4799]: > logger="UnhandledError" Nov 24 07:04:25 crc kubenswrapper[4799]: E1124 07:04:25.759827 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/ae46b5b6-4f76-4730-923e-7c44c7eaafbe/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" podUID="ae46b5b6-4f76-4730-923e-7c44c7eaafbe" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.085752 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n" event={"ID":"dd32bea6-7ad9-4182-bc59-5644efe24ce1","Type":"ContainerStarted","Data":"5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983"} Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.087679 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-2dz2n" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.095291 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"997ffb9e-cc6a-4fca-91d9-eb17d711b135","Type":"ContainerStarted","Data":"664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2"} Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.098900 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kgrrs" event={"ID":"a7742c81-45e0-4ef5-aceb-7a631bf9cc37","Type":"ContainerStarted","Data":"8bb8a5532f39845617e4497e4831243eb9b594226daca424f0f2fa01515b4c8b"} Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.104494 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"04342f4b-48ae-4c45-ad6b-947b30b9de75","Type":"ContainerStarted","Data":"7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c"} Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.104671 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.107005 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b0361baf-d0c3-430f-87e3-c011f869a2b3","Type":"ContainerStarted","Data":"4bcf19e3b710fbf877a92e707435d95962e55b4801e15a2e5b43f689d28b8bba"} Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.107108 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.112531 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-2dz2n" podStartSLOduration=10.219513848 podStartE2EDuration="17.112506981s" podCreationTimestamp="2025-11-24 07:04:09 +0000 UTC" firstStartedPulling="2025-11-24 07:04:18.533518775 +0000 UTC m=+1004.189501249" lastFinishedPulling="2025-11-24 07:04:25.426511888 +0000 UTC m=+1011.082494382" observedRunningTime="2025-11-24 07:04:26.109926018 +0000 UTC m=+1011.765908492" watchObservedRunningTime="2025-11-24 07:04:26.112506981 +0000 UTC m=+1011.768489455" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.121711 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace","Type":"ContainerStarted","Data":"477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742"} Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.129957 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bbfb60f9-6e71-4b22-9d74-cab607a00c20","Type":"ContainerStarted","Data":"7d643d949a4bb973305b6b986dbd4165fc4a1f0f0846279bc0b5ddf8929ad32b"} Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.134041 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"844d53db-a457-4153-bdb4-d557ba34c099","Type":"ContainerStarted","Data":"5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029"} Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.145283 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=13.831187184000001 podStartE2EDuration="21.145264248s" podCreationTimestamp="2025-11-24 07:04:05 +0000 UTC" firstStartedPulling="2025-11-24 07:04:18.200798833 +0000 UTC m=+1003.856781307" lastFinishedPulling="2025-11-24 07:04:25.514875897 +0000 UTC m=+1011.170858371" observedRunningTime="2025-11-24 07:04:26.134427551 +0000 UTC m=+1011.790410025" watchObservedRunningTime="2025-11-24 07:04:26.145264248 +0000 UTC m=+1011.801246722" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.166370 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=16.987934863 podStartE2EDuration="23.166347774s" podCreationTimestamp="2025-11-24 07:04:03 +0000 UTC" firstStartedPulling="2025-11-24 07:04:18.369049512 +0000 UTC m=+1004.025031986" lastFinishedPulling="2025-11-24 07:04:24.547462423 +0000 UTC m=+1010.203444897" observedRunningTime="2025-11-24 07:04:26.16511927 +0000 UTC m=+1011.821101744" watchObservedRunningTime="2025-11-24 07:04:26.166347774 +0000 UTC m=+1011.822330248" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.449173 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.579648 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-config\") pod \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.580141 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-dns-svc\") pod \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.580206 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm97x\" (UniqueName: \"kubernetes.io/projected/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-kube-api-access-nm97x\") pod \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\" (UID: \"ae46b5b6-4f76-4730-923e-7c44c7eaafbe\") " Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.588064 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-kube-api-access-nm97x" (OuterVolumeSpecName: "kube-api-access-nm97x") pod "ae46b5b6-4f76-4730-923e-7c44c7eaafbe" (UID: "ae46b5b6-4f76-4730-923e-7c44c7eaafbe"). InnerVolumeSpecName "kube-api-access-nm97x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.621198 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-config" (OuterVolumeSpecName: "config") pod "ae46b5b6-4f76-4730-923e-7c44c7eaafbe" (UID: "ae46b5b6-4f76-4730-923e-7c44c7eaafbe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.651976 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ae46b5b6-4f76-4730-923e-7c44c7eaafbe" (UID: "ae46b5b6-4f76-4730-923e-7c44c7eaafbe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.682019 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.682064 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm97x\" (UniqueName: \"kubernetes.io/projected/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-kube-api-access-nm97x\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:26 crc kubenswrapper[4799]: I1124 07:04:26.682077 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae46b5b6-4f76-4730-923e-7c44c7eaafbe-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:27 crc kubenswrapper[4799]: I1124 07:04:27.145935 4799 generic.go:334] "Generic (PLEG): container finished" podID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerID="8bb8a5532f39845617e4497e4831243eb9b594226daca424f0f2fa01515b4c8b" exitCode=0 Nov 24 07:04:27 crc kubenswrapper[4799]: I1124 07:04:27.146041 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kgrrs" event={"ID":"a7742c81-45e0-4ef5-aceb-7a631bf9cc37","Type":"ContainerDied","Data":"8bb8a5532f39845617e4497e4831243eb9b594226daca424f0f2fa01515b4c8b"} Nov 24 07:04:27 crc kubenswrapper[4799]: I1124 07:04:27.155816 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" Nov 24 07:04:27 crc kubenswrapper[4799]: I1124 07:04:27.157418 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-7np9d" event={"ID":"ae46b5b6-4f76-4730-923e-7c44c7eaafbe","Type":"ContainerDied","Data":"2c7e172f4f7929289607d7e0441e5bf7d8189638aaea82ce127cfa8a5c30a430"} Nov 24 07:04:27 crc kubenswrapper[4799]: I1124 07:04:27.157486 4799 scope.go:117] "RemoveContainer" containerID="a48042e0db19d83f9c784d1d6228c57c41f55f99545e98f19aef096cc18b2df7" Nov 24 07:04:27 crc kubenswrapper[4799]: I1124 07:04:27.220994 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-7np9d"] Nov 24 07:04:27 crc kubenswrapper[4799]: I1124 07:04:27.222894 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-7np9d"] Nov 24 07:04:27 crc kubenswrapper[4799]: I1124 07:04:27.642970 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae46b5b6-4f76-4730-923e-7c44c7eaafbe" path="/var/lib/kubelet/pods/ae46b5b6-4f76-4730-923e-7c44c7eaafbe/volumes" Nov 24 07:04:28 crc kubenswrapper[4799]: I1124 07:04:28.164996 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kgrrs" event={"ID":"a7742c81-45e0-4ef5-aceb-7a631bf9cc37","Type":"ContainerStarted","Data":"343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81"} Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.200661 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace","Type":"ContainerStarted","Data":"2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c"} Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.205101 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bbfb60f9-6e71-4b22-9d74-cab607a00c20","Type":"ContainerStarted","Data":"a9ca7d86df080d4b606da7e5a800c32f1929eb3f6ccde8a25eb20a206bdf9fda"} Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.208321 4799 generic.go:334] "Generic (PLEG): container finished" podID="844d53db-a457-4153-bdb4-d557ba34c099" containerID="5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029" exitCode=0 Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.208429 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"844d53db-a457-4153-bdb4-d557ba34c099","Type":"ContainerDied","Data":"5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029"} Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.214491 4799 generic.go:334] "Generic (PLEG): container finished" podID="997ffb9e-cc6a-4fca-91d9-eb17d711b135" containerID="664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2" exitCode=0 Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.214584 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"997ffb9e-cc6a-4fca-91d9-eb17d711b135","Type":"ContainerDied","Data":"664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2"} Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.221156 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kgrrs" event={"ID":"a7742c81-45e0-4ef5-aceb-7a631bf9cc37","Type":"ContainerStarted","Data":"16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f"} Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.221602 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.221804 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.234286 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=8.666226793 podStartE2EDuration="19.234266579s" podCreationTimestamp="2025-11-24 07:04:11 +0000 UTC" firstStartedPulling="2025-11-24 07:04:18.679264857 +0000 UTC m=+1004.335247331" lastFinishedPulling="2025-11-24 07:04:29.247304603 +0000 UTC m=+1014.903287117" observedRunningTime="2025-11-24 07:04:30.230495043 +0000 UTC m=+1015.886477517" watchObservedRunningTime="2025-11-24 07:04:30.234266579 +0000 UTC m=+1015.890249053" Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.297299 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=11.588815294 podStartE2EDuration="22.297280182s" podCreationTimestamp="2025-11-24 07:04:08 +0000 UTC" firstStartedPulling="2025-11-24 07:04:18.556411712 +0000 UTC m=+1004.212394186" lastFinishedPulling="2025-11-24 07:04:29.26487658 +0000 UTC m=+1014.920859074" observedRunningTime="2025-11-24 07:04:30.292798495 +0000 UTC m=+1015.948780969" watchObservedRunningTime="2025-11-24 07:04:30.297280182 +0000 UTC m=+1015.953262656" Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.346528 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-kgrrs" podStartSLOduration=15.730141481 podStartE2EDuration="21.346491934s" podCreationTimestamp="2025-11-24 07:04:09 +0000 UTC" firstStartedPulling="2025-11-24 07:04:19.819326194 +0000 UTC m=+1005.475308668" lastFinishedPulling="2025-11-24 07:04:25.435676647 +0000 UTC m=+1011.091659121" observedRunningTime="2025-11-24 07:04:30.316831125 +0000 UTC m=+1015.972813599" watchObservedRunningTime="2025-11-24 07:04:30.346491934 +0000 UTC m=+1016.002474418" Nov 24 07:04:30 crc kubenswrapper[4799]: I1124 07:04:30.956980 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.021219 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.066925 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.122069 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.237019 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"844d53db-a457-4153-bdb4-d557ba34c099","Type":"ContainerStarted","Data":"57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e"} Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.242722 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"997ffb9e-cc6a-4fca-91d9-eb17d711b135","Type":"ContainerStarted","Data":"63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174"} Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.243277 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.243698 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.277771 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=22.711979817 podStartE2EDuration="29.277737235s" podCreationTimestamp="2025-11-24 07:04:02 +0000 UTC" firstStartedPulling="2025-11-24 07:04:18.200453273 +0000 UTC m=+1003.856435757" lastFinishedPulling="2025-11-24 07:04:24.766210711 +0000 UTC m=+1010.422193175" observedRunningTime="2025-11-24 07:04:31.268154014 +0000 UTC m=+1016.924136488" watchObservedRunningTime="2025-11-24 07:04:31.277737235 +0000 UTC m=+1016.933719719" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.298895 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.302570 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=24.394014484 podStartE2EDuration="31.302541696s" podCreationTimestamp="2025-11-24 07:04:00 +0000 UTC" firstStartedPulling="2025-11-24 07:04:18.527449163 +0000 UTC m=+1004.183431637" lastFinishedPulling="2025-11-24 07:04:25.435976375 +0000 UTC m=+1011.091958849" observedRunningTime="2025-11-24 07:04:31.295503227 +0000 UTC m=+1016.951485741" watchObservedRunningTime="2025-11-24 07:04:31.302541696 +0000 UTC m=+1016.958524190" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.307523 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.530348 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c65c5f57f-n6m64"] Nov 24 07:04:31 crc kubenswrapper[4799]: E1124 07:04:31.530725 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae46b5b6-4f76-4730-923e-7c44c7eaafbe" containerName="init" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.530740 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae46b5b6-4f76-4730-923e-7c44c7eaafbe" containerName="init" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.530943 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae46b5b6-4f76-4730-923e-7c44c7eaafbe" containerName="init" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.531891 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.534660 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.546959 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-fxr57"] Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.548261 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.551665 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.560733 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c65c5f57f-n6m64"] Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.578322 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-fxr57"] Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.686150 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c65c5f57f-n6m64"] Nov 24 07:04:31 crc kubenswrapper[4799]: E1124 07:04:31.686892 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-5xb6g ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" podUID="912cfe60-e8ae-4c56-947b-00ca67a75493" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.701473 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovn-rundir\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.701575 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.701673 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-dns-svc\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.701705 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovs-rundir\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.701731 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-combined-ca-bundle\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.701761 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-config\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.701977 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flvrz\" (UniqueName: \"kubernetes.io/projected/09082d4e-0623-40c5-b584-1eb58985cfc0-kube-api-access-flvrz\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.702070 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09082d4e-0623-40c5-b584-1eb58985cfc0-config\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.702168 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-ovsdbserver-nb\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.702229 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xb6g\" (UniqueName: \"kubernetes.io/projected/912cfe60-e8ae-4c56-947b-00ca67a75493-kube-api-access-5xb6g\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.716945 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-g7w8k"] Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.720376 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.723651 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.733579 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-g7w8k"] Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.795882 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.797541 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.803026 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.803555 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-5nszq" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.803823 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.811405 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-ovsdbserver-nb\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.811689 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xb6g\" (UniqueName: \"kubernetes.io/projected/912cfe60-e8ae-4c56-947b-00ca67a75493-kube-api-access-5xb6g\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.811876 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovn-rundir\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.812045 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.812217 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-dns-svc\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.812343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovs-rundir\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.812460 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-combined-ca-bundle\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.812584 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-config\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.812719 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flvrz\" (UniqueName: \"kubernetes.io/projected/09082d4e-0623-40c5-b584-1eb58985cfc0-kube-api-access-flvrz\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.812833 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09082d4e-0623-40c5-b584-1eb58985cfc0-config\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.812907 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-ovsdbserver-nb\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.813955 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovs-rundir\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.813330 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.813187 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovn-rundir\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.814596 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-config\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.814765 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09082d4e-0623-40c5-b584-1eb58985cfc0-config\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.825750 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-dns-svc\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.826932 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-combined-ca-bundle\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.829069 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.841719 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.842453 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flvrz\" (UniqueName: \"kubernetes.io/projected/09082d4e-0623-40c5-b584-1eb58985cfc0-kube-api-access-flvrz\") pod \"ovn-controller-metrics-fxr57\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.842949 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xb6g\" (UniqueName: \"kubernetes.io/projected/912cfe60-e8ae-4c56-947b-00ca67a75493-kube-api-access-5xb6g\") pod \"dnsmasq-dns-6c65c5f57f-n6m64\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.870388 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.914993 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-config\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.915064 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.915093 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-config\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.915130 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.915169 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.916101 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.916247 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2rt8\" (UniqueName: \"kubernetes.io/projected/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-kube-api-access-d2rt8\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.916331 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.916412 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb55z\" (UniqueName: \"kubernetes.io/projected/6aa59428-ee60-4434-9245-3ff8fa55200e-kube-api-access-fb55z\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.916518 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.916613 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:31 crc kubenswrapper[4799]: I1124 07:04:31.916733 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-scripts\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.018674 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb55z\" (UniqueName: \"kubernetes.io/projected/6aa59428-ee60-4434-9245-3ff8fa55200e-kube-api-access-fb55z\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.018772 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.018800 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.018833 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-scripts\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.018930 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-config\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.018968 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.018998 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-config\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.019043 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.019088 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.019117 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.019155 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2rt8\" (UniqueName: \"kubernetes.io/projected/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-kube-api-access-d2rt8\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.019182 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.037189 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.038098 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-scripts\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.038797 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-config\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.039960 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.040104 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.040125 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-config\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.041432 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.046792 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.048051 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.048747 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.057669 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb55z\" (UniqueName: \"kubernetes.io/projected/6aa59428-ee60-4434-9245-3ff8fa55200e-kube-api-access-fb55z\") pod \"ovn-northd-0\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.062496 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2rt8\" (UniqueName: \"kubernetes.io/projected/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-kube-api-access-d2rt8\") pod \"dnsmasq-dns-5c476d78c5-g7w8k\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.197086 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.197126 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.254647 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.257900 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.266691 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.354727 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.420740 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-fxr57"] Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.424887 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-config\") pod \"912cfe60-e8ae-4c56-947b-00ca67a75493\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.425487 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xb6g\" (UniqueName: \"kubernetes.io/projected/912cfe60-e8ae-4c56-947b-00ca67a75493-kube-api-access-5xb6g\") pod \"912cfe60-e8ae-4c56-947b-00ca67a75493\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.425934 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-ovsdbserver-nb\") pod \"912cfe60-e8ae-4c56-947b-00ca67a75493\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.425435 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-config" (OuterVolumeSpecName: "config") pod "912cfe60-e8ae-4c56-947b-00ca67a75493" (UID: "912cfe60-e8ae-4c56-947b-00ca67a75493"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.426046 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-dns-svc\") pod \"912cfe60-e8ae-4c56-947b-00ca67a75493\" (UID: \"912cfe60-e8ae-4c56-947b-00ca67a75493\") " Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.426672 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "912cfe60-e8ae-4c56-947b-00ca67a75493" (UID: "912cfe60-e8ae-4c56-947b-00ca67a75493"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.427870 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "912cfe60-e8ae-4c56-947b-00ca67a75493" (UID: "912cfe60-e8ae-4c56-947b-00ca67a75493"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.428052 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.428072 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.433696 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/912cfe60-e8ae-4c56-947b-00ca67a75493-kube-api-access-5xb6g" (OuterVolumeSpecName: "kube-api-access-5xb6g") pod "912cfe60-e8ae-4c56-947b-00ca67a75493" (UID: "912cfe60-e8ae-4c56-947b-00ca67a75493"). InnerVolumeSpecName "kube-api-access-5xb6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.530000 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/912cfe60-e8ae-4c56-947b-00ca67a75493-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.530274 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xb6g\" (UniqueName: \"kubernetes.io/projected/912cfe60-e8ae-4c56-947b-00ca67a75493-kube-api-access-5xb6g\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.755972 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 07:04:32 crc kubenswrapper[4799]: W1124 07:04:32.759115 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6aa59428_ee60_4434_9245_3ff8fa55200e.slice/crio-975d6d84ae0a753e4c63ead344e1f49709456a53d737fca911099c1d0005019c WatchSource:0}: Error finding container 975d6d84ae0a753e4c63ead344e1f49709456a53d737fca911099c1d0005019c: Status 404 returned error can't find the container with id 975d6d84ae0a753e4c63ead344e1f49709456a53d737fca911099c1d0005019c Nov 24 07:04:32 crc kubenswrapper[4799]: I1124 07:04:32.876430 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-g7w8k"] Nov 24 07:04:32 crc kubenswrapper[4799]: W1124 07:04:32.880366 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8959b0ce_7cf8_4110_9dfd_8b83aed9b472.slice/crio-f4f8fc171b74de92501e42fd29aea322c4fee1bb52f1c05f029e29f58827363d WatchSource:0}: Error finding container f4f8fc171b74de92501e42fd29aea322c4fee1bb52f1c05f029e29f58827363d: Status 404 returned error can't find the container with id f4f8fc171b74de92501e42fd29aea322c4fee1bb52f1c05f029e29f58827363d Nov 24 07:04:33 crc kubenswrapper[4799]: I1124 07:04:33.262482 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6aa59428-ee60-4434-9245-3ff8fa55200e","Type":"ContainerStarted","Data":"975d6d84ae0a753e4c63ead344e1f49709456a53d737fca911099c1d0005019c"} Nov 24 07:04:33 crc kubenswrapper[4799]: I1124 07:04:33.263398 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" event={"ID":"8959b0ce-7cf8-4110-9dfd-8b83aed9b472","Type":"ContainerStarted","Data":"f4f8fc171b74de92501e42fd29aea322c4fee1bb52f1c05f029e29f58827363d"} Nov 24 07:04:33 crc kubenswrapper[4799]: I1124 07:04:33.264656 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fxr57" event={"ID":"09082d4e-0623-40c5-b584-1eb58985cfc0","Type":"ContainerStarted","Data":"cbe00e7f26e6d398423c98a60db1c3f591d35a88157f4ac3ab54a88e3f4aaaf4"} Nov 24 07:04:33 crc kubenswrapper[4799]: I1124 07:04:33.264729 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c65c5f57f-n6m64" Nov 24 07:04:33 crc kubenswrapper[4799]: I1124 07:04:33.363770 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c65c5f57f-n6m64"] Nov 24 07:04:33 crc kubenswrapper[4799]: I1124 07:04:33.369686 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c65c5f57f-n6m64"] Nov 24 07:04:33 crc kubenswrapper[4799]: I1124 07:04:33.611002 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:33 crc kubenswrapper[4799]: I1124 07:04:33.611079 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:33 crc kubenswrapper[4799]: I1124 07:04:33.646730 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="912cfe60-e8ae-4c56-947b-00ca67a75493" path="/var/lib/kubelet/pods/912cfe60-e8ae-4c56-947b-00ca67a75493/volumes" Nov 24 07:04:33 crc kubenswrapper[4799]: I1124 07:04:33.806050 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 07:04:35 crc kubenswrapper[4799]: I1124 07:04:35.878385 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.058954 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-g7w8k"] Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.089962 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-94d98"] Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.091896 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.101275 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-94d98"] Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.220549 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.220632 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6lj4\" (UniqueName: \"kubernetes.io/projected/9774b0e9-c82a-4c51-aa43-251b9220dae3-kube-api-access-b6lj4\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.220780 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-config\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.220938 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.221203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-dns-svc\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.322685 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.322742 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6lj4\" (UniqueName: \"kubernetes.io/projected/9774b0e9-c82a-4c51-aa43-251b9220dae3-kube-api-access-b6lj4\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.322815 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-config\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.322882 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.322928 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-dns-svc\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.323809 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.323827 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-dns-svc\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.323902 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-config\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.324083 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.343656 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6lj4\" (UniqueName: \"kubernetes.io/projected/9774b0e9-c82a-4c51-aa43-251b9220dae3-kube-api-access-b6lj4\") pod \"dnsmasq-dns-5c9fdb784c-94d98\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.426804 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:36 crc kubenswrapper[4799]: I1124 07:04:36.903587 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-94d98"] Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.179078 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.187225 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.191096 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.191549 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.195917 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.196579 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-9b9cb" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.218661 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.297408 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" event={"ID":"9774b0e9-c82a-4c51-aa43-251b9220dae3","Type":"ContainerStarted","Data":"f38b3ff57bf7c89677d8186f97f02dae73bc1814f026471b8fec822a1cb4e1ab"} Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.356460 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckrjg\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-kube-api-access-ckrjg\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.356669 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-cache\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.356756 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-lock\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.357129 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.357342 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.459396 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckrjg\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-kube-api-access-ckrjg\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.459536 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-cache\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.459617 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-lock\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.459735 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.459908 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.460544 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.461096 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-lock\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: E1124 07:04:37.461553 4799 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 07:04:37 crc kubenswrapper[4799]: E1124 07:04:37.461592 4799 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 07:04:37 crc kubenswrapper[4799]: E1124 07:04:37.461660 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift podName:a30589ef-eff4-4113-8e6a-32f03be5c734 nodeName:}" failed. No retries permitted until 2025-11-24 07:04:37.961638771 +0000 UTC m=+1023.617621325 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift") pod "swift-storage-0" (UID: "a30589ef-eff4-4113-8e6a-32f03be5c734") : configmap "swift-ring-files" not found Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.461673 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-cache\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.487655 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckrjg\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-kube-api-access-ckrjg\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.505216 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.684550 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-d57xl"] Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.686963 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.696597 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.696667 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.696722 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.710839 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-d57xl"] Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.766106 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-scripts\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.766251 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-dispersionconf\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.766308 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-ring-data-devices\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.766403 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-combined-ca-bundle\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.766442 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2f82\" (UniqueName: \"kubernetes.io/projected/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-kube-api-access-v2f82\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.766473 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-swiftconf\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.766540 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-etc-swift\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.867810 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-scripts\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.867931 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-dispersionconf\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.867969 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-ring-data-devices\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.868032 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-combined-ca-bundle\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.868064 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2f82\" (UniqueName: \"kubernetes.io/projected/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-kube-api-access-v2f82\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.868093 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-swiftconf\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.868143 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-etc-swift\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.868594 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-etc-swift\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.868767 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-scripts\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.869321 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-ring-data-devices\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.874275 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-dispersionconf\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.878733 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-swiftconf\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.880346 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-combined-ca-bundle\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.902005 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2f82\" (UniqueName: \"kubernetes.io/projected/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-kube-api-access-v2f82\") pod \"swift-ring-rebalance-d57xl\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:37 crc kubenswrapper[4799]: I1124 07:04:37.969628 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:37 crc kubenswrapper[4799]: E1124 07:04:37.969802 4799 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 07:04:37 crc kubenswrapper[4799]: E1124 07:04:37.969822 4799 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 07:04:37 crc kubenswrapper[4799]: E1124 07:04:37.969963 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift podName:a30589ef-eff4-4113-8e6a-32f03be5c734 nodeName:}" failed. No retries permitted until 2025-11-24 07:04:38.969872087 +0000 UTC m=+1024.625854561 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift") pod "swift-storage-0" (UID: "a30589ef-eff4-4113-8e6a-32f03be5c734") : configmap "swift-ring-files" not found Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.014821 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.278272 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.323020 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fxr57" event={"ID":"09082d4e-0623-40c5-b584-1eb58985cfc0","Type":"ContainerStarted","Data":"145fcceaa41ce0287f7a6fe79891e468626f0454f5b1fca70bfabb879bb1c5fa"} Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.329866 4799 generic.go:334] "Generic (PLEG): container finished" podID="9774b0e9-c82a-4c51-aa43-251b9220dae3" containerID="847abe5903aab2744f125f425787c7b139e9c7c11ce8a965535b10689fc42575" exitCode=0 Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.329922 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" event={"ID":"9774b0e9-c82a-4c51-aa43-251b9220dae3","Type":"ContainerDied","Data":"847abe5903aab2744f125f425787c7b139e9c7c11ce8a965535b10689fc42575"} Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.332787 4799 generic.go:334] "Generic (PLEG): container finished" podID="8959b0ce-7cf8-4110-9dfd-8b83aed9b472" containerID="4c39db05a17ec5ff720a4e6f14e0145da80bd1cdc757cd694f7136485737a940" exitCode=0 Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.332822 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" event={"ID":"8959b0ce-7cf8-4110-9dfd-8b83aed9b472","Type":"ContainerDied","Data":"4c39db05a17ec5ff720a4e6f14e0145da80bd1cdc757cd694f7136485737a940"} Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.346173 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-fxr57" podStartSLOduration=7.346156421 podStartE2EDuration="7.346156421s" podCreationTimestamp="2025-11-24 07:04:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:04:38.338502734 +0000 UTC m=+1023.994485208" watchObservedRunningTime="2025-11-24 07:04:38.346156421 +0000 UTC m=+1024.002138895" Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.489602 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.492373 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-d57xl"] Nov 24 07:04:38 crc kubenswrapper[4799]: W1124 07:04:38.833654 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98cd3086_ad5c_4ca7_bcf7_e68e52ef8053.slice/crio-8d6518615d3554b9082d377ca1e469ab0130cb20d75367e6fde59397cb7f3062 WatchSource:0}: Error finding container 8d6518615d3554b9082d377ca1e469ab0130cb20d75367e6fde59397cb7f3062: Status 404 returned error can't find the container with id 8d6518615d3554b9082d377ca1e469ab0130cb20d75367e6fde59397cb7f3062 Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.982743 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:38 crc kubenswrapper[4799]: I1124 07:04:38.991536 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:38 crc kubenswrapper[4799]: E1124 07:04:38.991787 4799 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 07:04:38 crc kubenswrapper[4799]: E1124 07:04:38.991811 4799 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 07:04:38 crc kubenswrapper[4799]: E1124 07:04:38.991902 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift podName:a30589ef-eff4-4113-8e6a-32f03be5c734 nodeName:}" failed. No retries permitted until 2025-11-24 07:04:40.991882026 +0000 UTC m=+1026.647864510 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift") pod "swift-storage-0" (UID: "a30589ef-eff4-4113-8e6a-32f03be5c734") : configmap "swift-ring-files" not found Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.094648 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2rt8\" (UniqueName: \"kubernetes.io/projected/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-kube-api-access-d2rt8\") pod \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.094865 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-config\") pod \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.094979 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-sb\") pod \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.095035 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-nb\") pod \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.095064 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-dns-svc\") pod \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\" (UID: \"8959b0ce-7cf8-4110-9dfd-8b83aed9b472\") " Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.101801 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-kube-api-access-d2rt8" (OuterVolumeSpecName: "kube-api-access-d2rt8") pod "8959b0ce-7cf8-4110-9dfd-8b83aed9b472" (UID: "8959b0ce-7cf8-4110-9dfd-8b83aed9b472"). InnerVolumeSpecName "kube-api-access-d2rt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.120815 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8959b0ce-7cf8-4110-9dfd-8b83aed9b472" (UID: "8959b0ce-7cf8-4110-9dfd-8b83aed9b472"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.128696 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8959b0ce-7cf8-4110-9dfd-8b83aed9b472" (UID: "8959b0ce-7cf8-4110-9dfd-8b83aed9b472"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.131692 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8959b0ce-7cf8-4110-9dfd-8b83aed9b472" (UID: "8959b0ce-7cf8-4110-9dfd-8b83aed9b472"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.133641 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-config" (OuterVolumeSpecName: "config") pod "8959b0ce-7cf8-4110-9dfd-8b83aed9b472" (UID: "8959b0ce-7cf8-4110-9dfd-8b83aed9b472"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.200125 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.200167 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.200184 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.200196 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.200210 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2rt8\" (UniqueName: \"kubernetes.io/projected/8959b0ce-7cf8-4110-9dfd-8b83aed9b472-kube-api-access-d2rt8\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.351511 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" event={"ID":"9774b0e9-c82a-4c51-aa43-251b9220dae3","Type":"ContainerStarted","Data":"32cdd4ab29d85774618fc3ce4f3b80cf07da822841725b1d95407f40133e58a9"} Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.352885 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.363783 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d57xl" event={"ID":"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053","Type":"ContainerStarted","Data":"8d6518615d3554b9082d377ca1e469ab0130cb20d75367e6fde59397cb7f3062"} Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.371995 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.371992 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-g7w8k" event={"ID":"8959b0ce-7cf8-4110-9dfd-8b83aed9b472","Type":"ContainerDied","Data":"f4f8fc171b74de92501e42fd29aea322c4fee1bb52f1c05f029e29f58827363d"} Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.372273 4799 scope.go:117] "RemoveContainer" containerID="4c39db05a17ec5ff720a4e6f14e0145da80bd1cdc757cd694f7136485737a940" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.392104 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" podStartSLOduration=3.392054585 podStartE2EDuration="3.392054585s" podCreationTimestamp="2025-11-24 07:04:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:04:39.379093439 +0000 UTC m=+1025.035075913" watchObservedRunningTime="2025-11-24 07:04:39.392054585 +0000 UTC m=+1025.048037059" Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.577294 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-g7w8k"] Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.585015 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-g7w8k"] Nov 24 07:04:39 crc kubenswrapper[4799]: I1124 07:04:39.649137 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8959b0ce-7cf8-4110-9dfd-8b83aed9b472" path="/var/lib/kubelet/pods/8959b0ce-7cf8-4110-9dfd-8b83aed9b472/volumes" Nov 24 07:04:40 crc kubenswrapper[4799]: I1124 07:04:40.381547 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6aa59428-ee60-4434-9245-3ff8fa55200e","Type":"ContainerStarted","Data":"8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80"} Nov 24 07:04:40 crc kubenswrapper[4799]: I1124 07:04:40.381913 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6aa59428-ee60-4434-9245-3ff8fa55200e","Type":"ContainerStarted","Data":"2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1"} Nov 24 07:04:40 crc kubenswrapper[4799]: I1124 07:04:40.382628 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 07:04:40 crc kubenswrapper[4799]: I1124 07:04:40.407477 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.9990587189999998 podStartE2EDuration="9.407455256s" podCreationTimestamp="2025-11-24 07:04:31 +0000 UTC" firstStartedPulling="2025-11-24 07:04:32.763219733 +0000 UTC m=+1018.419202197" lastFinishedPulling="2025-11-24 07:04:39.17161626 +0000 UTC m=+1024.827598734" observedRunningTime="2025-11-24 07:04:40.401742684 +0000 UTC m=+1026.057725178" watchObservedRunningTime="2025-11-24 07:04:40.407455256 +0000 UTC m=+1026.063437730" Nov 24 07:04:41 crc kubenswrapper[4799]: I1124 07:04:41.040468 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:41 crc kubenswrapper[4799]: E1124 07:04:41.040656 4799 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 07:04:41 crc kubenswrapper[4799]: E1124 07:04:41.040683 4799 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 07:04:41 crc kubenswrapper[4799]: E1124 07:04:41.040744 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift podName:a30589ef-eff4-4113-8e6a-32f03be5c734 nodeName:}" failed. No retries permitted until 2025-11-24 07:04:45.040727979 +0000 UTC m=+1030.696710453 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift") pod "swift-storage-0" (UID: "a30589ef-eff4-4113-8e6a-32f03be5c734") : configmap "swift-ring-files" not found Nov 24 07:04:42 crc kubenswrapper[4799]: I1124 07:04:42.335415 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 07:04:42 crc kubenswrapper[4799]: I1124 07:04:42.406517 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d57xl" event={"ID":"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053","Type":"ContainerStarted","Data":"458ae0656978a384ca17474f208a13aed37fa1be14fa7cdbd76f5ef2136e7ccc"} Nov 24 07:04:42 crc kubenswrapper[4799]: I1124 07:04:42.433375 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-d57xl" podStartSLOduration=2.397393696 podStartE2EDuration="5.43334965s" podCreationTimestamp="2025-11-24 07:04:37 +0000 UTC" firstStartedPulling="2025-11-24 07:04:38.835729659 +0000 UTC m=+1024.491712133" lastFinishedPulling="2025-11-24 07:04:41.871685613 +0000 UTC m=+1027.527668087" observedRunningTime="2025-11-24 07:04:42.427455894 +0000 UTC m=+1028.083438378" watchObservedRunningTime="2025-11-24 07:04:42.43334965 +0000 UTC m=+1028.089332124" Nov 24 07:04:42 crc kubenswrapper[4799]: I1124 07:04:42.472284 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.481597 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cabe-account-create-zjzwn"] Nov 24 07:04:43 crc kubenswrapper[4799]: E1124 07:04:43.482004 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8959b0ce-7cf8-4110-9dfd-8b83aed9b472" containerName="init" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.482020 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8959b0ce-7cf8-4110-9dfd-8b83aed9b472" containerName="init" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.482195 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8959b0ce-7cf8-4110-9dfd-8b83aed9b472" containerName="init" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.482772 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cabe-account-create-zjzwn" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.485791 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.509722 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cabe-account-create-zjzwn"] Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.531773 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-ljnx2"] Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.532944 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ljnx2" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.548257 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ljnx2"] Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.591950 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32a13216-4d25-462a-b55b-59602988c3c4-operator-scripts\") pod \"keystone-cabe-account-create-zjzwn\" (UID: \"32a13216-4d25-462a-b55b-59602988c3c4\") " pod="openstack/keystone-cabe-account-create-zjzwn" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.592066 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98d23df1-e416-488c-b1bc-d075d5b63fb3-operator-scripts\") pod \"keystone-db-create-ljnx2\" (UID: \"98d23df1-e416-488c-b1bc-d075d5b63fb3\") " pod="openstack/keystone-db-create-ljnx2" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.592101 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4r2r\" (UniqueName: \"kubernetes.io/projected/32a13216-4d25-462a-b55b-59602988c3c4-kube-api-access-z4r2r\") pod \"keystone-cabe-account-create-zjzwn\" (UID: \"32a13216-4d25-462a-b55b-59602988c3c4\") " pod="openstack/keystone-cabe-account-create-zjzwn" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.592159 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw2fn\" (UniqueName: \"kubernetes.io/projected/98d23df1-e416-488c-b1bc-d075d5b63fb3-kube-api-access-tw2fn\") pod \"keystone-db-create-ljnx2\" (UID: \"98d23df1-e416-488c-b1bc-d075d5b63fb3\") " pod="openstack/keystone-db-create-ljnx2" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.702006 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32a13216-4d25-462a-b55b-59602988c3c4-operator-scripts\") pod \"keystone-cabe-account-create-zjzwn\" (UID: \"32a13216-4d25-462a-b55b-59602988c3c4\") " pod="openstack/keystone-cabe-account-create-zjzwn" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.702172 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98d23df1-e416-488c-b1bc-d075d5b63fb3-operator-scripts\") pod \"keystone-db-create-ljnx2\" (UID: \"98d23df1-e416-488c-b1bc-d075d5b63fb3\") " pod="openstack/keystone-db-create-ljnx2" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.702275 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4r2r\" (UniqueName: \"kubernetes.io/projected/32a13216-4d25-462a-b55b-59602988c3c4-kube-api-access-z4r2r\") pod \"keystone-cabe-account-create-zjzwn\" (UID: \"32a13216-4d25-462a-b55b-59602988c3c4\") " pod="openstack/keystone-cabe-account-create-zjzwn" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.702496 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw2fn\" (UniqueName: \"kubernetes.io/projected/98d23df1-e416-488c-b1bc-d075d5b63fb3-kube-api-access-tw2fn\") pod \"keystone-db-create-ljnx2\" (UID: \"98d23df1-e416-488c-b1bc-d075d5b63fb3\") " pod="openstack/keystone-db-create-ljnx2" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.705751 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98d23df1-e416-488c-b1bc-d075d5b63fb3-operator-scripts\") pod \"keystone-db-create-ljnx2\" (UID: \"98d23df1-e416-488c-b1bc-d075d5b63fb3\") " pod="openstack/keystone-db-create-ljnx2" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.708892 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32a13216-4d25-462a-b55b-59602988c3c4-operator-scripts\") pod \"keystone-cabe-account-create-zjzwn\" (UID: \"32a13216-4d25-462a-b55b-59602988c3c4\") " pod="openstack/keystone-cabe-account-create-zjzwn" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.717515 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-wsc8j"] Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.718923 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wsc8j" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.735924 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wsc8j"] Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.751776 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw2fn\" (UniqueName: \"kubernetes.io/projected/98d23df1-e416-488c-b1bc-d075d5b63fb3-kube-api-access-tw2fn\") pod \"keystone-db-create-ljnx2\" (UID: \"98d23df1-e416-488c-b1bc-d075d5b63fb3\") " pod="openstack/keystone-db-create-ljnx2" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.765382 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4r2r\" (UniqueName: \"kubernetes.io/projected/32a13216-4d25-462a-b55b-59602988c3c4-kube-api-access-z4r2r\") pod \"keystone-cabe-account-create-zjzwn\" (UID: \"32a13216-4d25-462a-b55b-59602988c3c4\") " pod="openstack/keystone-cabe-account-create-zjzwn" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.805763 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-operator-scripts\") pod \"placement-db-create-wsc8j\" (UID: \"d7128a3b-75e5-49c4-b25c-8d2a643d65cb\") " pod="openstack/placement-db-create-wsc8j" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.806045 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7cnm\" (UniqueName: \"kubernetes.io/projected/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-kube-api-access-h7cnm\") pod \"placement-db-create-wsc8j\" (UID: \"d7128a3b-75e5-49c4-b25c-8d2a643d65cb\") " pod="openstack/placement-db-create-wsc8j" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.809632 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cabe-account-create-zjzwn" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.812660 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-2704-account-create-mt2ps"] Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.814159 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2704-account-create-mt2ps" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.818498 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.842697 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2704-account-create-mt2ps"] Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.858525 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ljnx2" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.908839 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-operator-scripts\") pod \"placement-db-create-wsc8j\" (UID: \"d7128a3b-75e5-49c4-b25c-8d2a643d65cb\") " pod="openstack/placement-db-create-wsc8j" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.908959 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7cnm\" (UniqueName: \"kubernetes.io/projected/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-kube-api-access-h7cnm\") pod \"placement-db-create-wsc8j\" (UID: \"d7128a3b-75e5-49c4-b25c-8d2a643d65cb\") " pod="openstack/placement-db-create-wsc8j" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.909147 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3d5dc9f-307c-4217-a7df-c1b6551441a5-operator-scripts\") pod \"placement-2704-account-create-mt2ps\" (UID: \"c3d5dc9f-307c-4217-a7df-c1b6551441a5\") " pod="openstack/placement-2704-account-create-mt2ps" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.909225 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f9ch\" (UniqueName: \"kubernetes.io/projected/c3d5dc9f-307c-4217-a7df-c1b6551441a5-kube-api-access-4f9ch\") pod \"placement-2704-account-create-mt2ps\" (UID: \"c3d5dc9f-307c-4217-a7df-c1b6551441a5\") " pod="openstack/placement-2704-account-create-mt2ps" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.910150 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-operator-scripts\") pod \"placement-db-create-wsc8j\" (UID: \"d7128a3b-75e5-49c4-b25c-8d2a643d65cb\") " pod="openstack/placement-db-create-wsc8j" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.940084 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7cnm\" (UniqueName: \"kubernetes.io/projected/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-kube-api-access-h7cnm\") pod \"placement-db-create-wsc8j\" (UID: \"d7128a3b-75e5-49c4-b25c-8d2a643d65cb\") " pod="openstack/placement-db-create-wsc8j" Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.976117 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-b5qgf"] Nov 24 07:04:43 crc kubenswrapper[4799]: I1124 07:04:43.978864 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-b5qgf" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.011764 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3d5dc9f-307c-4217-a7df-c1b6551441a5-operator-scripts\") pod \"placement-2704-account-create-mt2ps\" (UID: \"c3d5dc9f-307c-4217-a7df-c1b6551441a5\") " pod="openstack/placement-2704-account-create-mt2ps" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.011814 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f9ch\" (UniqueName: \"kubernetes.io/projected/c3d5dc9f-307c-4217-a7df-c1b6551441a5-kube-api-access-4f9ch\") pod \"placement-2704-account-create-mt2ps\" (UID: \"c3d5dc9f-307c-4217-a7df-c1b6551441a5\") " pod="openstack/placement-2704-account-create-mt2ps" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.012735 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3d5dc9f-307c-4217-a7df-c1b6551441a5-operator-scripts\") pod \"placement-2704-account-create-mt2ps\" (UID: \"c3d5dc9f-307c-4217-a7df-c1b6551441a5\") " pod="openstack/placement-2704-account-create-mt2ps" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.071419 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f9ch\" (UniqueName: \"kubernetes.io/projected/c3d5dc9f-307c-4217-a7df-c1b6551441a5-kube-api-access-4f9ch\") pod \"placement-2704-account-create-mt2ps\" (UID: \"c3d5dc9f-307c-4217-a7df-c1b6551441a5\") " pod="openstack/placement-2704-account-create-mt2ps" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.116866 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-operator-scripts\") pod \"glance-db-create-b5qgf\" (UID: \"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e\") " pod="openstack/glance-db-create-b5qgf" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.117031 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghfw6\" (UniqueName: \"kubernetes.io/projected/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-kube-api-access-ghfw6\") pod \"glance-db-create-b5qgf\" (UID: \"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e\") " pod="openstack/glance-db-create-b5qgf" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.117377 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wsc8j" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.053228 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-b5qgf"] Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.188913 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-00d0-account-create-xqldv"] Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.190145 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-00d0-account-create-xqldv" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.192641 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.197924 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-00d0-account-create-xqldv"] Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.222274 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-operator-scripts\") pod \"glance-db-create-b5qgf\" (UID: \"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e\") " pod="openstack/glance-db-create-b5qgf" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.222378 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzghz\" (UniqueName: \"kubernetes.io/projected/644e0295-5410-47b5-8cb0-ffc0f89d8b13-kube-api-access-nzghz\") pod \"glance-00d0-account-create-xqldv\" (UID: \"644e0295-5410-47b5-8cb0-ffc0f89d8b13\") " pod="openstack/glance-00d0-account-create-xqldv" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.222424 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghfw6\" (UniqueName: \"kubernetes.io/projected/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-kube-api-access-ghfw6\") pod \"glance-db-create-b5qgf\" (UID: \"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e\") " pod="openstack/glance-db-create-b5qgf" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.222473 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/644e0295-5410-47b5-8cb0-ffc0f89d8b13-operator-scripts\") pod \"glance-00d0-account-create-xqldv\" (UID: \"644e0295-5410-47b5-8cb0-ffc0f89d8b13\") " pod="openstack/glance-00d0-account-create-xqldv" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.223588 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-operator-scripts\") pod \"glance-db-create-b5qgf\" (UID: \"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e\") " pod="openstack/glance-db-create-b5qgf" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.241566 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2704-account-create-mt2ps" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.268158 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghfw6\" (UniqueName: \"kubernetes.io/projected/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-kube-api-access-ghfw6\") pod \"glance-db-create-b5qgf\" (UID: \"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e\") " pod="openstack/glance-db-create-b5qgf" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.331441 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzghz\" (UniqueName: \"kubernetes.io/projected/644e0295-5410-47b5-8cb0-ffc0f89d8b13-kube-api-access-nzghz\") pod \"glance-00d0-account-create-xqldv\" (UID: \"644e0295-5410-47b5-8cb0-ffc0f89d8b13\") " pod="openstack/glance-00d0-account-create-xqldv" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.331953 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/644e0295-5410-47b5-8cb0-ffc0f89d8b13-operator-scripts\") pod \"glance-00d0-account-create-xqldv\" (UID: \"644e0295-5410-47b5-8cb0-ffc0f89d8b13\") " pod="openstack/glance-00d0-account-create-xqldv" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.332639 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/644e0295-5410-47b5-8cb0-ffc0f89d8b13-operator-scripts\") pod \"glance-00d0-account-create-xqldv\" (UID: \"644e0295-5410-47b5-8cb0-ffc0f89d8b13\") " pod="openstack/glance-00d0-account-create-xqldv" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.348593 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzghz\" (UniqueName: \"kubernetes.io/projected/644e0295-5410-47b5-8cb0-ffc0f89d8b13-kube-api-access-nzghz\") pod \"glance-00d0-account-create-xqldv\" (UID: \"644e0295-5410-47b5-8cb0-ffc0f89d8b13\") " pod="openstack/glance-00d0-account-create-xqldv" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.373325 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-b5qgf" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.562171 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-00d0-account-create-xqldv" Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.597166 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cabe-account-create-zjzwn"] Nov 24 07:04:44 crc kubenswrapper[4799]: W1124 07:04:44.599967 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98d23df1_e416_488c_b1bc_d075d5b63fb3.slice/crio-b00abf8a22a42b192d0b9bd2bbe9ea1e8261cc1b6bea5de923e8fa596d16b3d0 WatchSource:0}: Error finding container b00abf8a22a42b192d0b9bd2bbe9ea1e8261cc1b6bea5de923e8fa596d16b3d0: Status 404 returned error can't find the container with id b00abf8a22a42b192d0b9bd2bbe9ea1e8261cc1b6bea5de923e8fa596d16b3d0 Nov 24 07:04:44 crc kubenswrapper[4799]: W1124 07:04:44.603743 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32a13216_4d25_462a_b55b_59602988c3c4.slice/crio-83ed81e2f00544eb977974c897ee455b1da5e478d3d65b81343a06a02f0b32e9 WatchSource:0}: Error finding container 83ed81e2f00544eb977974c897ee455b1da5e478d3d65b81343a06a02f0b32e9: Status 404 returned error can't find the container with id 83ed81e2f00544eb977974c897ee455b1da5e478d3d65b81343a06a02f0b32e9 Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.605052 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ljnx2"] Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.849465 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wsc8j"] Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.863745 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2704-account-create-mt2ps"] Nov 24 07:04:44 crc kubenswrapper[4799]: I1124 07:04:44.945128 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-b5qgf"] Nov 24 07:04:44 crc kubenswrapper[4799]: W1124 07:04:44.971248 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9e7a595_81bd_4b15_8682_fbdf8ab08e5e.slice/crio-5df8421ce3118c80bd3b5d074fd852fe2adf10de24b54bcc2a9c03a30918af47 WatchSource:0}: Error finding container 5df8421ce3118c80bd3b5d074fd852fe2adf10de24b54bcc2a9c03a30918af47: Status 404 returned error can't find the container with id 5df8421ce3118c80bd3b5d074fd852fe2adf10de24b54bcc2a9c03a30918af47 Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.051964 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:45 crc kubenswrapper[4799]: E1124 07:04:45.052180 4799 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 07:04:45 crc kubenswrapper[4799]: E1124 07:04:45.052210 4799 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 07:04:45 crc kubenswrapper[4799]: E1124 07:04:45.052274 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift podName:a30589ef-eff4-4113-8e6a-32f03be5c734 nodeName:}" failed. No retries permitted until 2025-11-24 07:04:53.052252548 +0000 UTC m=+1038.708235022 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift") pod "swift-storage-0" (UID: "a30589ef-eff4-4113-8e6a-32f03be5c734") : configmap "swift-ring-files" not found Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.102184 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-00d0-account-create-xqldv"] Nov 24 07:04:45 crc kubenswrapper[4799]: W1124 07:04:45.178972 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod644e0295_5410_47b5_8cb0_ffc0f89d8b13.slice/crio-0f9f0260e6de2864aabc4172f9b64ce945f01984a39bdf68b845f4781b325d6f WatchSource:0}: Error finding container 0f9f0260e6de2864aabc4172f9b64ce945f01984a39bdf68b845f4781b325d6f: Status 404 returned error can't find the container with id 0f9f0260e6de2864aabc4172f9b64ce945f01984a39bdf68b845f4781b325d6f Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.444900 4799 generic.go:334] "Generic (PLEG): container finished" podID="d7128a3b-75e5-49c4-b25c-8d2a643d65cb" containerID="fd71cd5d765ff4cf61393c1642ef7e752e23da9d8a6512420626cd2e2713c9a3" exitCode=0 Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.444970 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wsc8j" event={"ID":"d7128a3b-75e5-49c4-b25c-8d2a643d65cb","Type":"ContainerDied","Data":"fd71cd5d765ff4cf61393c1642ef7e752e23da9d8a6512420626cd2e2713c9a3"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.445003 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wsc8j" event={"ID":"d7128a3b-75e5-49c4-b25c-8d2a643d65cb","Type":"ContainerStarted","Data":"e87688fe8061ff6adc58d1f7868086c3b72d9aa54ba412cb6d30a866a0b80e9b"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.449409 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-00d0-account-create-xqldv" event={"ID":"644e0295-5410-47b5-8cb0-ffc0f89d8b13","Type":"ContainerStarted","Data":"e8c89665cca133cc14d7182c955b9f158d183661351a87c4fa15352e6936b60e"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.449506 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-00d0-account-create-xqldv" event={"ID":"644e0295-5410-47b5-8cb0-ffc0f89d8b13","Type":"ContainerStarted","Data":"0f9f0260e6de2864aabc4172f9b64ce945f01984a39bdf68b845f4781b325d6f"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.452148 4799 generic.go:334] "Generic (PLEG): container finished" podID="98d23df1-e416-488c-b1bc-d075d5b63fb3" containerID="ffedbe819747f9260bf81488c2e8b318f2f357ada47ae430c3886e5eda43a1fd" exitCode=0 Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.452212 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ljnx2" event={"ID":"98d23df1-e416-488c-b1bc-d075d5b63fb3","Type":"ContainerDied","Data":"ffedbe819747f9260bf81488c2e8b318f2f357ada47ae430c3886e5eda43a1fd"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.452240 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ljnx2" event={"ID":"98d23df1-e416-488c-b1bc-d075d5b63fb3","Type":"ContainerStarted","Data":"b00abf8a22a42b192d0b9bd2bbe9ea1e8261cc1b6bea5de923e8fa596d16b3d0"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.454227 4799 generic.go:334] "Generic (PLEG): container finished" podID="a9e7a595-81bd-4b15-8682-fbdf8ab08e5e" containerID="e4394c9753dd8146da5af402052bb93b4dc93f3348696e0f8ee006cd1df32c5c" exitCode=0 Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.454271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-b5qgf" event={"ID":"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e","Type":"ContainerDied","Data":"e4394c9753dd8146da5af402052bb93b4dc93f3348696e0f8ee006cd1df32c5c"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.454291 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-b5qgf" event={"ID":"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e","Type":"ContainerStarted","Data":"5df8421ce3118c80bd3b5d074fd852fe2adf10de24b54bcc2a9c03a30918af47"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.455877 4799 generic.go:334] "Generic (PLEG): container finished" podID="c3d5dc9f-307c-4217-a7df-c1b6551441a5" containerID="6c441a620e5b48c41c3cf1709a813e69634cba1f9b860df79644b13ccfc40292" exitCode=0 Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.455924 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2704-account-create-mt2ps" event={"ID":"c3d5dc9f-307c-4217-a7df-c1b6551441a5","Type":"ContainerDied","Data":"6c441a620e5b48c41c3cf1709a813e69634cba1f9b860df79644b13ccfc40292"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.455972 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2704-account-create-mt2ps" event={"ID":"c3d5dc9f-307c-4217-a7df-c1b6551441a5","Type":"ContainerStarted","Data":"340ec691ead37b3adf8b40ba70520a9fc8e026a6200822ad943116195ebfc3f8"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.457650 4799 generic.go:334] "Generic (PLEG): container finished" podID="32a13216-4d25-462a-b55b-59602988c3c4" containerID="c54fcecc5a5d26755e67a7a6d3e539af8f51f5fc6dd58f3a088edd47fd54d01d" exitCode=0 Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.457699 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cabe-account-create-zjzwn" event={"ID":"32a13216-4d25-462a-b55b-59602988c3c4","Type":"ContainerDied","Data":"c54fcecc5a5d26755e67a7a6d3e539af8f51f5fc6dd58f3a088edd47fd54d01d"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.457741 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cabe-account-create-zjzwn" event={"ID":"32a13216-4d25-462a-b55b-59602988c3c4","Type":"ContainerStarted","Data":"83ed81e2f00544eb977974c897ee455b1da5e478d3d65b81343a06a02f0b32e9"} Nov 24 07:04:45 crc kubenswrapper[4799]: I1124 07:04:45.525231 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-00d0-account-create-xqldv" podStartSLOduration=1.525214596 podStartE2EDuration="1.525214596s" podCreationTimestamp="2025-11-24 07:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:04:45.520219855 +0000 UTC m=+1031.176202339" watchObservedRunningTime="2025-11-24 07:04:45.525214596 +0000 UTC m=+1031.181197070" Nov 24 07:04:46 crc kubenswrapper[4799]: I1124 07:04:46.428121 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:04:46 crc kubenswrapper[4799]: I1124 07:04:46.475197 4799 generic.go:334] "Generic (PLEG): container finished" podID="644e0295-5410-47b5-8cb0-ffc0f89d8b13" containerID="e8c89665cca133cc14d7182c955b9f158d183661351a87c4fa15352e6936b60e" exitCode=0 Nov 24 07:04:46 crc kubenswrapper[4799]: I1124 07:04:46.475931 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-00d0-account-create-xqldv" event={"ID":"644e0295-5410-47b5-8cb0-ffc0f89d8b13","Type":"ContainerDied","Data":"e8c89665cca133cc14d7182c955b9f158d183661351a87c4fa15352e6936b60e"} Nov 24 07:04:46 crc kubenswrapper[4799]: I1124 07:04:46.512933 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-h94gz"] Nov 24 07:04:46 crc kubenswrapper[4799]: I1124 07:04:46.513233 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" podUID="0285d647-978a-4e9b-8ba2-b369a038db04" containerName="dnsmasq-dns" containerID="cri-o://0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338" gracePeriod=10 Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.073793 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2704-account-create-mt2ps" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.078223 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wsc8j" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.099453 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-operator-scripts\") pod \"d7128a3b-75e5-49c4-b25c-8d2a643d65cb\" (UID: \"d7128a3b-75e5-49c4-b25c-8d2a643d65cb\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.099516 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7cnm\" (UniqueName: \"kubernetes.io/projected/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-kube-api-access-h7cnm\") pod \"d7128a3b-75e5-49c4-b25c-8d2a643d65cb\" (UID: \"d7128a3b-75e5-49c4-b25c-8d2a643d65cb\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.099698 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f9ch\" (UniqueName: \"kubernetes.io/projected/c3d5dc9f-307c-4217-a7df-c1b6551441a5-kube-api-access-4f9ch\") pod \"c3d5dc9f-307c-4217-a7df-c1b6551441a5\" (UID: \"c3d5dc9f-307c-4217-a7df-c1b6551441a5\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.099763 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3d5dc9f-307c-4217-a7df-c1b6551441a5-operator-scripts\") pod \"c3d5dc9f-307c-4217-a7df-c1b6551441a5\" (UID: \"c3d5dc9f-307c-4217-a7df-c1b6551441a5\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.100787 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7128a3b-75e5-49c4-b25c-8d2a643d65cb" (UID: "d7128a3b-75e5-49c4-b25c-8d2a643d65cb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.100813 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3d5dc9f-307c-4217-a7df-c1b6551441a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c3d5dc9f-307c-4217-a7df-c1b6551441a5" (UID: "c3d5dc9f-307c-4217-a7df-c1b6551441a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.120080 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3d5dc9f-307c-4217-a7df-c1b6551441a5-kube-api-access-4f9ch" (OuterVolumeSpecName: "kube-api-access-4f9ch") pod "c3d5dc9f-307c-4217-a7df-c1b6551441a5" (UID: "c3d5dc9f-307c-4217-a7df-c1b6551441a5"). InnerVolumeSpecName "kube-api-access-4f9ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.123585 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-kube-api-access-h7cnm" (OuterVolumeSpecName: "kube-api-access-h7cnm") pod "d7128a3b-75e5-49c4-b25c-8d2a643d65cb" (UID: "d7128a3b-75e5-49c4-b25c-8d2a643d65cb"). InnerVolumeSpecName "kube-api-access-h7cnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.211412 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f9ch\" (UniqueName: \"kubernetes.io/projected/c3d5dc9f-307c-4217-a7df-c1b6551441a5-kube-api-access-4f9ch\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.211470 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3d5dc9f-307c-4217-a7df-c1b6551441a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.211482 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.211492 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7cnm\" (UniqueName: \"kubernetes.io/projected/d7128a3b-75e5-49c4-b25c-8d2a643d65cb-kube-api-access-h7cnm\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.354204 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.398076 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cabe-account-create-zjzwn" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.429057 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-config\") pod \"0285d647-978a-4e9b-8ba2-b369a038db04\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.429125 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzb2m\" (UniqueName: \"kubernetes.io/projected/0285d647-978a-4e9b-8ba2-b369a038db04-kube-api-access-vzb2m\") pod \"0285d647-978a-4e9b-8ba2-b369a038db04\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.429253 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-dns-svc\") pod \"0285d647-978a-4e9b-8ba2-b369a038db04\" (UID: \"0285d647-978a-4e9b-8ba2-b369a038db04\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.440459 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0285d647-978a-4e9b-8ba2-b369a038db04-kube-api-access-vzb2m" (OuterVolumeSpecName: "kube-api-access-vzb2m") pod "0285d647-978a-4e9b-8ba2-b369a038db04" (UID: "0285d647-978a-4e9b-8ba2-b369a038db04"). InnerVolumeSpecName "kube-api-access-vzb2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.500554 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ljnx2" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.502097 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-b5qgf" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.505785 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cabe-account-create-zjzwn" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.505804 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cabe-account-create-zjzwn" event={"ID":"32a13216-4d25-462a-b55b-59602988c3c4","Type":"ContainerDied","Data":"83ed81e2f00544eb977974c897ee455b1da5e478d3d65b81343a06a02f0b32e9"} Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.506110 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83ed81e2f00544eb977974c897ee455b1da5e478d3d65b81343a06a02f0b32e9" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.506415 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-config" (OuterVolumeSpecName: "config") pod "0285d647-978a-4e9b-8ba2-b369a038db04" (UID: "0285d647-978a-4e9b-8ba2-b369a038db04"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.518058 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wsc8j" event={"ID":"d7128a3b-75e5-49c4-b25c-8d2a643d65cb","Type":"ContainerDied","Data":"e87688fe8061ff6adc58d1f7868086c3b72d9aa54ba412cb6d30a866a0b80e9b"} Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.518098 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e87688fe8061ff6adc58d1f7868086c3b72d9aa54ba412cb6d30a866a0b80e9b" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.518169 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wsc8j" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.532813 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32a13216-4d25-462a-b55b-59602988c3c4-operator-scripts\") pod \"32a13216-4d25-462a-b55b-59602988c3c4\" (UID: \"32a13216-4d25-462a-b55b-59602988c3c4\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.533205 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4r2r\" (UniqueName: \"kubernetes.io/projected/32a13216-4d25-462a-b55b-59602988c3c4-kube-api-access-z4r2r\") pod \"32a13216-4d25-462a-b55b-59602988c3c4\" (UID: \"32a13216-4d25-462a-b55b-59602988c3c4\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.534386 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.534418 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzb2m\" (UniqueName: \"kubernetes.io/projected/0285d647-978a-4e9b-8ba2-b369a038db04-kube-api-access-vzb2m\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.534643 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32a13216-4d25-462a-b55b-59602988c3c4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32a13216-4d25-462a-b55b-59602988c3c4" (UID: "32a13216-4d25-462a-b55b-59602988c3c4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.535009 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-b5qgf" event={"ID":"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e","Type":"ContainerDied","Data":"5df8421ce3118c80bd3b5d074fd852fe2adf10de24b54bcc2a9c03a30918af47"} Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.535048 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5df8421ce3118c80bd3b5d074fd852fe2adf10de24b54bcc2a9c03a30918af47" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.535102 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-b5qgf" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.537665 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32a13216-4d25-462a-b55b-59602988c3c4-kube-api-access-z4r2r" (OuterVolumeSpecName: "kube-api-access-z4r2r") pod "32a13216-4d25-462a-b55b-59602988c3c4" (UID: "32a13216-4d25-462a-b55b-59602988c3c4"). InnerVolumeSpecName "kube-api-access-z4r2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.539279 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ljnx2" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.539299 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ljnx2" event={"ID":"98d23df1-e416-488c-b1bc-d075d5b63fb3","Type":"ContainerDied","Data":"b00abf8a22a42b192d0b9bd2bbe9ea1e8261cc1b6bea5de923e8fa596d16b3d0"} Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.539346 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b00abf8a22a42b192d0b9bd2bbe9ea1e8261cc1b6bea5de923e8fa596d16b3d0" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.542211 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2704-account-create-mt2ps" event={"ID":"c3d5dc9f-307c-4217-a7df-c1b6551441a5","Type":"ContainerDied","Data":"340ec691ead37b3adf8b40ba70520a9fc8e026a6200822ad943116195ebfc3f8"} Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.542255 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="340ec691ead37b3adf8b40ba70520a9fc8e026a6200822ad943116195ebfc3f8" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.542295 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2704-account-create-mt2ps" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.566438 4799 generic.go:334] "Generic (PLEG): container finished" podID="0285d647-978a-4e9b-8ba2-b369a038db04" containerID="0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338" exitCode=0 Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.566644 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.567534 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" event={"ID":"0285d647-978a-4e9b-8ba2-b369a038db04","Type":"ContainerDied","Data":"0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338"} Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.567610 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-h94gz" event={"ID":"0285d647-978a-4e9b-8ba2-b369a038db04","Type":"ContainerDied","Data":"c4563b2aaf818bc7d31b548f44b49b4f543d70d369fe9a30546f0694edca1853"} Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.567679 4799 scope.go:117] "RemoveContainer" containerID="0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.568897 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0285d647-978a-4e9b-8ba2-b369a038db04" (UID: "0285d647-978a-4e9b-8ba2-b369a038db04"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.621864 4799 scope.go:117] "RemoveContainer" containerID="5044f3331cc3cd478cb2dce11a0fec48a73f5eeccf2e4bbc37bf383f6c48534b" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.635347 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-operator-scripts\") pod \"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e\" (UID: \"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.635484 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98d23df1-e416-488c-b1bc-d075d5b63fb3-operator-scripts\") pod \"98d23df1-e416-488c-b1bc-d075d5b63fb3\" (UID: \"98d23df1-e416-488c-b1bc-d075d5b63fb3\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.635596 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw2fn\" (UniqueName: \"kubernetes.io/projected/98d23df1-e416-488c-b1bc-d075d5b63fb3-kube-api-access-tw2fn\") pod \"98d23df1-e416-488c-b1bc-d075d5b63fb3\" (UID: \"98d23df1-e416-488c-b1bc-d075d5b63fb3\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.635662 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghfw6\" (UniqueName: \"kubernetes.io/projected/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-kube-api-access-ghfw6\") pod \"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e\" (UID: \"a9e7a595-81bd-4b15-8682-fbdf8ab08e5e\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.635817 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a9e7a595-81bd-4b15-8682-fbdf8ab08e5e" (UID: "a9e7a595-81bd-4b15-8682-fbdf8ab08e5e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.636100 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.636113 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32a13216-4d25-462a-b55b-59602988c3c4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.636124 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4r2r\" (UniqueName: \"kubernetes.io/projected/32a13216-4d25-462a-b55b-59602988c3c4-kube-api-access-z4r2r\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.636137 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0285d647-978a-4e9b-8ba2-b369a038db04-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.636339 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98d23df1-e416-488c-b1bc-d075d5b63fb3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "98d23df1-e416-488c-b1bc-d075d5b63fb3" (UID: "98d23df1-e416-488c-b1bc-d075d5b63fb3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.641572 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-kube-api-access-ghfw6" (OuterVolumeSpecName: "kube-api-access-ghfw6") pod "a9e7a595-81bd-4b15-8682-fbdf8ab08e5e" (UID: "a9e7a595-81bd-4b15-8682-fbdf8ab08e5e"). InnerVolumeSpecName "kube-api-access-ghfw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.641623 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98d23df1-e416-488c-b1bc-d075d5b63fb3-kube-api-access-tw2fn" (OuterVolumeSpecName: "kube-api-access-tw2fn") pod "98d23df1-e416-488c-b1bc-d075d5b63fb3" (UID: "98d23df1-e416-488c-b1bc-d075d5b63fb3"). InnerVolumeSpecName "kube-api-access-tw2fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.645151 4799 scope.go:117] "RemoveContainer" containerID="0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338" Nov 24 07:04:47 crc kubenswrapper[4799]: E1124 07:04:47.649277 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338\": container with ID starting with 0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338 not found: ID does not exist" containerID="0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.649325 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338"} err="failed to get container status \"0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338\": rpc error: code = NotFound desc = could not find container \"0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338\": container with ID starting with 0266872e56064988808505c9fc4bbcd807b0b90ab336091918316c9f5aef3338 not found: ID does not exist" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.649379 4799 scope.go:117] "RemoveContainer" containerID="5044f3331cc3cd478cb2dce11a0fec48a73f5eeccf2e4bbc37bf383f6c48534b" Nov 24 07:04:47 crc kubenswrapper[4799]: E1124 07:04:47.652808 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5044f3331cc3cd478cb2dce11a0fec48a73f5eeccf2e4bbc37bf383f6c48534b\": container with ID starting with 5044f3331cc3cd478cb2dce11a0fec48a73f5eeccf2e4bbc37bf383f6c48534b not found: ID does not exist" containerID="5044f3331cc3cd478cb2dce11a0fec48a73f5eeccf2e4bbc37bf383f6c48534b" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.652880 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5044f3331cc3cd478cb2dce11a0fec48a73f5eeccf2e4bbc37bf383f6c48534b"} err="failed to get container status \"5044f3331cc3cd478cb2dce11a0fec48a73f5eeccf2e4bbc37bf383f6c48534b\": rpc error: code = NotFound desc = could not find container \"5044f3331cc3cd478cb2dce11a0fec48a73f5eeccf2e4bbc37bf383f6c48534b\": container with ID starting with 5044f3331cc3cd478cb2dce11a0fec48a73f5eeccf2e4bbc37bf383f6c48534b not found: ID does not exist" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.737897 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw2fn\" (UniqueName: \"kubernetes.io/projected/98d23df1-e416-488c-b1bc-d075d5b63fb3-kube-api-access-tw2fn\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.737933 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghfw6\" (UniqueName: \"kubernetes.io/projected/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e-kube-api-access-ghfw6\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.737943 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98d23df1-e416-488c-b1bc-d075d5b63fb3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.861599 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-00d0-account-create-xqldv" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.915319 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-h94gz"] Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.920997 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-h94gz"] Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.942193 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/644e0295-5410-47b5-8cb0-ffc0f89d8b13-operator-scripts\") pod \"644e0295-5410-47b5-8cb0-ffc0f89d8b13\" (UID: \"644e0295-5410-47b5-8cb0-ffc0f89d8b13\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.942263 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzghz\" (UniqueName: \"kubernetes.io/projected/644e0295-5410-47b5-8cb0-ffc0f89d8b13-kube-api-access-nzghz\") pod \"644e0295-5410-47b5-8cb0-ffc0f89d8b13\" (UID: \"644e0295-5410-47b5-8cb0-ffc0f89d8b13\") " Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.942928 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/644e0295-5410-47b5-8cb0-ffc0f89d8b13-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "644e0295-5410-47b5-8cb0-ffc0f89d8b13" (UID: "644e0295-5410-47b5-8cb0-ffc0f89d8b13"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:47 crc kubenswrapper[4799]: I1124 07:04:47.945992 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/644e0295-5410-47b5-8cb0-ffc0f89d8b13-kube-api-access-nzghz" (OuterVolumeSpecName: "kube-api-access-nzghz") pod "644e0295-5410-47b5-8cb0-ffc0f89d8b13" (UID: "644e0295-5410-47b5-8cb0-ffc0f89d8b13"). InnerVolumeSpecName "kube-api-access-nzghz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:48 crc kubenswrapper[4799]: I1124 07:04:48.045027 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/644e0295-5410-47b5-8cb0-ffc0f89d8b13-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:48 crc kubenswrapper[4799]: I1124 07:04:48.045083 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzghz\" (UniqueName: \"kubernetes.io/projected/644e0295-5410-47b5-8cb0-ffc0f89d8b13-kube-api-access-nzghz\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:48 crc kubenswrapper[4799]: I1124 07:04:48.580763 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-00d0-account-create-xqldv" event={"ID":"644e0295-5410-47b5-8cb0-ffc0f89d8b13","Type":"ContainerDied","Data":"0f9f0260e6de2864aabc4172f9b64ce945f01984a39bdf68b845f4781b325d6f"} Nov 24 07:04:48 crc kubenswrapper[4799]: I1124 07:04:48.580826 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f9f0260e6de2864aabc4172f9b64ce945f01984a39bdf68b845f4781b325d6f" Nov 24 07:04:48 crc kubenswrapper[4799]: I1124 07:04:48.580914 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-00d0-account-create-xqldv" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.229964 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-5wg47"] Nov 24 07:04:49 crc kubenswrapper[4799]: E1124 07:04:49.230933 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7128a3b-75e5-49c4-b25c-8d2a643d65cb" containerName="mariadb-database-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.230959 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7128a3b-75e5-49c4-b25c-8d2a643d65cb" containerName="mariadb-database-create" Nov 24 07:04:49 crc kubenswrapper[4799]: E1124 07:04:49.230990 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="644e0295-5410-47b5-8cb0-ffc0f89d8b13" containerName="mariadb-account-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231000 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="644e0295-5410-47b5-8cb0-ffc0f89d8b13" containerName="mariadb-account-create" Nov 24 07:04:49 crc kubenswrapper[4799]: E1124 07:04:49.231026 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9e7a595-81bd-4b15-8682-fbdf8ab08e5e" containerName="mariadb-database-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231035 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9e7a595-81bd-4b15-8682-fbdf8ab08e5e" containerName="mariadb-database-create" Nov 24 07:04:49 crc kubenswrapper[4799]: E1124 07:04:49.231051 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98d23df1-e416-488c-b1bc-d075d5b63fb3" containerName="mariadb-database-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231059 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="98d23df1-e416-488c-b1bc-d075d5b63fb3" containerName="mariadb-database-create" Nov 24 07:04:49 crc kubenswrapper[4799]: E1124 07:04:49.231077 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32a13216-4d25-462a-b55b-59602988c3c4" containerName="mariadb-account-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231085 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="32a13216-4d25-462a-b55b-59602988c3c4" containerName="mariadb-account-create" Nov 24 07:04:49 crc kubenswrapper[4799]: E1124 07:04:49.231102 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3d5dc9f-307c-4217-a7df-c1b6551441a5" containerName="mariadb-account-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231110 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3d5dc9f-307c-4217-a7df-c1b6551441a5" containerName="mariadb-account-create" Nov 24 07:04:49 crc kubenswrapper[4799]: E1124 07:04:49.231130 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0285d647-978a-4e9b-8ba2-b369a038db04" containerName="dnsmasq-dns" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231137 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0285d647-978a-4e9b-8ba2-b369a038db04" containerName="dnsmasq-dns" Nov 24 07:04:49 crc kubenswrapper[4799]: E1124 07:04:49.231159 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0285d647-978a-4e9b-8ba2-b369a038db04" containerName="init" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231167 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0285d647-978a-4e9b-8ba2-b369a038db04" containerName="init" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231584 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="32a13216-4d25-462a-b55b-59602988c3c4" containerName="mariadb-account-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231663 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9e7a595-81bd-4b15-8682-fbdf8ab08e5e" containerName="mariadb-database-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231695 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="644e0295-5410-47b5-8cb0-ffc0f89d8b13" containerName="mariadb-account-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231722 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7128a3b-75e5-49c4-b25c-8d2a643d65cb" containerName="mariadb-database-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231764 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="98d23df1-e416-488c-b1bc-d075d5b63fb3" containerName="mariadb-database-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231781 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0285d647-978a-4e9b-8ba2-b369a038db04" containerName="dnsmasq-dns" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.231831 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3d5dc9f-307c-4217-a7df-c1b6551441a5" containerName="mariadb-account-create" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.232969 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.236783 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.236957 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kg5dw" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.284537 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-5wg47"] Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.371012 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-db-sync-config-data\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.371123 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgdjm\" (UniqueName: \"kubernetes.io/projected/03546c3d-2da5-4476-8c22-5a14939cfd80-kube-api-access-bgdjm\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.371217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-combined-ca-bundle\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.371245 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-config-data\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.472709 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-combined-ca-bundle\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.473115 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-config-data\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.473461 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-db-sync-config-data\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.473602 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgdjm\" (UniqueName: \"kubernetes.io/projected/03546c3d-2da5-4476-8c22-5a14939cfd80-kube-api-access-bgdjm\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.478862 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-db-sync-config-data\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.487246 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-config-data\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.487789 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-combined-ca-bundle\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.491111 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgdjm\" (UniqueName: \"kubernetes.io/projected/03546c3d-2da5-4476-8c22-5a14939cfd80-kube-api-access-bgdjm\") pod \"glance-db-sync-5wg47\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.565283 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5wg47" Nov 24 07:04:49 crc kubenswrapper[4799]: I1124 07:04:49.642104 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0285d647-978a-4e9b-8ba2-b369a038db04" path="/var/lib/kubelet/pods/0285d647-978a-4e9b-8ba2-b369a038db04/volumes" Nov 24 07:04:50 crc kubenswrapper[4799]: I1124 07:04:50.131822 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-5wg47"] Nov 24 07:04:50 crc kubenswrapper[4799]: W1124 07:04:50.166356 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03546c3d_2da5_4476_8c22_5a14939cfd80.slice/crio-f57d8fd5133b831ae3570505d1e8e07f7c1a52dbf6e18b8f97414bb752636596 WatchSource:0}: Error finding container f57d8fd5133b831ae3570505d1e8e07f7c1a52dbf6e18b8f97414bb752636596: Status 404 returned error can't find the container with id f57d8fd5133b831ae3570505d1e8e07f7c1a52dbf6e18b8f97414bb752636596 Nov 24 07:04:50 crc kubenswrapper[4799]: I1124 07:04:50.602113 4799 generic.go:334] "Generic (PLEG): container finished" podID="98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" containerID="458ae0656978a384ca17474f208a13aed37fa1be14fa7cdbd76f5ef2136e7ccc" exitCode=0 Nov 24 07:04:50 crc kubenswrapper[4799]: I1124 07:04:50.602213 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d57xl" event={"ID":"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053","Type":"ContainerDied","Data":"458ae0656978a384ca17474f208a13aed37fa1be14fa7cdbd76f5ef2136e7ccc"} Nov 24 07:04:50 crc kubenswrapper[4799]: I1124 07:04:50.606471 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5wg47" event={"ID":"03546c3d-2da5-4476-8c22-5a14939cfd80","Type":"ContainerStarted","Data":"f57d8fd5133b831ae3570505d1e8e07f7c1a52dbf6e18b8f97414bb752636596"} Nov 24 07:04:51 crc kubenswrapper[4799]: I1124 07:04:51.621680 4799 generic.go:334] "Generic (PLEG): container finished" podID="96acf88b-3fe8-46ac-8393-37f720632dd6" containerID="0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29" exitCode=0 Nov 24 07:04:51 crc kubenswrapper[4799]: I1124 07:04:51.621746 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"96acf88b-3fe8-46ac-8393-37f720632dd6","Type":"ContainerDied","Data":"0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29"} Nov 24 07:04:51 crc kubenswrapper[4799]: I1124 07:04:51.624944 4799 generic.go:334] "Generic (PLEG): container finished" podID="366b7548-2fca-4623-a45c-c1e0367ce93a" containerID="c6e22b9dd219757023e664a05027fc4af7ee03f3c7e62fe2a381ed00aef63570" exitCode=0 Nov 24 07:04:51 crc kubenswrapper[4799]: I1124 07:04:51.625223 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"366b7548-2fca-4623-a45c-c1e0367ce93a","Type":"ContainerDied","Data":"c6e22b9dd219757023e664a05027fc4af7ee03f3c7e62fe2a381ed00aef63570"} Nov 24 07:04:51 crc kubenswrapper[4799]: I1124 07:04:51.953559 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.026121 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-combined-ca-bundle\") pod \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.026246 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-ring-data-devices\") pod \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.026440 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-etc-swift\") pod \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.026487 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-dispersionconf\") pod \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.026584 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-scripts\") pod \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.026610 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2f82\" (UniqueName: \"kubernetes.io/projected/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-kube-api-access-v2f82\") pod \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.026671 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-swiftconf\") pod \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\" (UID: \"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053\") " Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.035510 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" (UID: "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.036796 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" (UID: "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.048685 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-kube-api-access-v2f82" (OuterVolumeSpecName: "kube-api-access-v2f82") pod "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" (UID: "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053"). InnerVolumeSpecName "kube-api-access-v2f82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.075639 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" (UID: "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.080404 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-scripts" (OuterVolumeSpecName: "scripts") pod "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" (UID: "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.084802 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" (UID: "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.089373 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" (UID: "98cd3086-ad5c-4ca7-bcf7-e68e52ef8053"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.129239 4799 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.129281 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.129295 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2f82\" (UniqueName: \"kubernetes.io/projected/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-kube-api-access-v2f82\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.129311 4799 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.129321 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.129332 4799 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.129344 4799 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.331441 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.639236 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"366b7548-2fca-4623-a45c-c1e0367ce93a","Type":"ContainerStarted","Data":"54abd42b7590aea7f2db6c595fca4f645386857fcbdb38711dfb864deb083510"} Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.639451 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.646543 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d57xl" event={"ID":"98cd3086-ad5c-4ca7-bcf7-e68e52ef8053","Type":"ContainerDied","Data":"8d6518615d3554b9082d377ca1e469ab0130cb20d75367e6fde59397cb7f3062"} Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.646585 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d6518615d3554b9082d377ca1e469ab0130cb20d75367e6fde59397cb7f3062" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.646676 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d57xl" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.669796 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"96acf88b-3fe8-46ac-8393-37f720632dd6","Type":"ContainerStarted","Data":"2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca"} Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.672100 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.682300 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.608917022 podStartE2EDuration="53.682274869s" podCreationTimestamp="2025-11-24 07:03:59 +0000 UTC" firstStartedPulling="2025-11-24 07:04:01.63970549 +0000 UTC m=+987.295687974" lastFinishedPulling="2025-11-24 07:04:17.713063347 +0000 UTC m=+1003.369045821" observedRunningTime="2025-11-24 07:04:52.672981076 +0000 UTC m=+1038.328963570" watchObservedRunningTime="2025-11-24 07:04:52.682274869 +0000 UTC m=+1038.338257353" Nov 24 07:04:52 crc kubenswrapper[4799]: I1124 07:04:52.715230 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.264332855 podStartE2EDuration="53.715209471s" podCreationTimestamp="2025-11-24 07:03:59 +0000 UTC" firstStartedPulling="2025-11-24 07:04:01.299711743 +0000 UTC m=+986.955694217" lastFinishedPulling="2025-11-24 07:04:17.750588359 +0000 UTC m=+1003.406570833" observedRunningTime="2025-11-24 07:04:52.706234987 +0000 UTC m=+1038.362217451" watchObservedRunningTime="2025-11-24 07:04:52.715209471 +0000 UTC m=+1038.371191935" Nov 24 07:04:53 crc kubenswrapper[4799]: I1124 07:04:53.146134 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:53 crc kubenswrapper[4799]: I1124 07:04:53.163134 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift\") pod \"swift-storage-0\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " pod="openstack/swift-storage-0" Nov 24 07:04:53 crc kubenswrapper[4799]: I1124 07:04:53.455181 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 07:04:54 crc kubenswrapper[4799]: I1124 07:04:54.232858 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 07:04:54 crc kubenswrapper[4799]: W1124 07:04:54.242597 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda30589ef_eff4_4113_8e6a_32f03be5c734.slice/crio-c599b8c2be169b387750d307815794693bda4ae021f7c3c39ac452e297437fc2 WatchSource:0}: Error finding container c599b8c2be169b387750d307815794693bda4ae021f7c3c39ac452e297437fc2: Status 404 returned error can't find the container with id c599b8c2be169b387750d307815794693bda4ae021f7c3c39ac452e297437fc2 Nov 24 07:04:54 crc kubenswrapper[4799]: I1124 07:04:54.694928 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"c599b8c2be169b387750d307815794693bda4ae021f7c3c39ac452e297437fc2"} Nov 24 07:04:55 crc kubenswrapper[4799]: I1124 07:04:55.709340 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"3328a1af0097e404c16240f7713c0a9b7963f642d250b69998ef457311035c34"} Nov 24 07:04:55 crc kubenswrapper[4799]: I1124 07:04:55.709669 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"dec54396850ed098f8d988491f08fcf315975d08d63a8517e087677c141683a7"} Nov 24 07:04:56 crc kubenswrapper[4799]: I1124 07:04:56.727057 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"de17e7e1d402f63433a3438f6c8d6c44f1b24885c6c84bac90d88ef38ef8f5f7"} Nov 24 07:04:56 crc kubenswrapper[4799]: I1124 07:04:56.727121 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"a412226a89e3fd5d662ed68efbfbea22971efde4e4219cd6b56cf841e130fc92"} Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.245380 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-2dz2n" podUID="dd32bea6-7ad9-4182-bc59-5644efe24ce1" containerName="ovn-controller" probeResult="failure" output=< Nov 24 07:05:00 crc kubenswrapper[4799]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 07:05:00 crc kubenswrapper[4799]: > Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.266360 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.267216 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.538186 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2dz2n-config-bgp5x"] Nov 24 07:05:00 crc kubenswrapper[4799]: E1124 07:05:00.538637 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" containerName="swift-ring-rebalance" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.538661 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" containerName="swift-ring-rebalance" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.539554 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" containerName="swift-ring-rebalance" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.540566 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.546459 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.553822 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2dz2n-config-bgp5x"] Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.695977 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-log-ovn\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.696040 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.696062 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run-ovn\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.696128 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-additional-scripts\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.696208 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-scripts\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.696246 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttbt8\" (UniqueName: \"kubernetes.io/projected/31f1da8c-b203-4427-91cf-08024dbf6f52-kube-api-access-ttbt8\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.798192 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-log-ovn\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.798279 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.798307 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run-ovn\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.798373 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-additional-scripts\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.798490 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-scripts\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.798521 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttbt8\" (UniqueName: \"kubernetes.io/projected/31f1da8c-b203-4427-91cf-08024dbf6f52-kube-api-access-ttbt8\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.798639 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run-ovn\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.798639 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.798635 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-log-ovn\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.799347 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-additional-scripts\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.801160 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-scripts\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.831964 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttbt8\" (UniqueName: \"kubernetes.io/projected/31f1da8c-b203-4427-91cf-08024dbf6f52-kube-api-access-ttbt8\") pod \"ovn-controller-2dz2n-config-bgp5x\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:00 crc kubenswrapper[4799]: I1124 07:05:00.867934 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:01 crc kubenswrapper[4799]: I1124 07:05:01.049221 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="366b7548-2fca-4623-a45c-c1e0367ce93a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.105:5671: connect: connection refused" Nov 24 07:05:05 crc kubenswrapper[4799]: I1124 07:05:05.269403 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-2dz2n" podUID="dd32bea6-7ad9-4182-bc59-5644efe24ce1" containerName="ovn-controller" probeResult="failure" output=< Nov 24 07:05:05 crc kubenswrapper[4799]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 07:05:05 crc kubenswrapper[4799]: > Nov 24 07:05:08 crc kubenswrapper[4799]: E1124 07:05:08.977908 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:26bd7b0bd6070856aefef6fe754c547d55c056396ea30d879d34c2d49b5a1d29" Nov 24 07:05:08 crc kubenswrapper[4799]: E1124 07:05:08.978528 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:26bd7b0bd6070856aefef6fe754c547d55c056396ea30d879d34c2d49b5a1d29,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bgdjm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-5wg47_openstack(03546c3d-2da5-4476-8c22-5a14939cfd80): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:05:08 crc kubenswrapper[4799]: E1124 07:05:08.979895 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-5wg47" podUID="03546c3d-2da5-4476-8c22-5a14939cfd80" Nov 24 07:05:09 crc kubenswrapper[4799]: E1124 07:05:09.148608 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75" Nov 24 07:05:09 crc kubenswrapper[4799]: E1124 07:05:09.148783 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-server,Image:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75,Command:[/usr/bin/swift-container-server /etc/swift/container-server.conf.d -v],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:container,HostPort:0,ContainerPort:6201,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b7h56h9dh94h67bh697h95h55hbh555h556h675h5fdh57dh579h5fbh64fh5c9h687hb6h678h5d4h549h54h98h8ch564h5bh5bch55dhc8hf8q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:swift,ReadOnly:false,MountPath:/srv/node/pv,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-swift,ReadOnly:false,MountPath:/etc/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cache,ReadOnly:false,MountPath:/var/cache/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:lock,ReadOnly:false,MountPath:/var/lock,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ckrjg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42445,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-storage-0_openstack(a30589ef-eff4-4113-8e6a-32f03be5c734): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:05:09 crc kubenswrapper[4799]: I1124 07:05:09.407247 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2dz2n-config-bgp5x"] Nov 24 07:05:09 crc kubenswrapper[4799]: I1124 07:05:09.897975 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n-config-bgp5x" event={"ID":"31f1da8c-b203-4427-91cf-08024dbf6f52","Type":"ContainerStarted","Data":"bb57102415efc322a8bcd1bc7aa2a2615715511ee00dafa287e4f8bb115a5c1a"} Nov 24 07:05:09 crc kubenswrapper[4799]: I1124 07:05:09.898507 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n-config-bgp5x" event={"ID":"31f1da8c-b203-4427-91cf-08024dbf6f52","Type":"ContainerStarted","Data":"3d105a895e4f10cac21184e257510d6b1183cc0c0f542a1b1b935ca293f7d8bf"} Nov 24 07:05:09 crc kubenswrapper[4799]: E1124 07:05:09.899414 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:26bd7b0bd6070856aefef6fe754c547d55c056396ea30d879d34c2d49b5a1d29\\\"\"" pod="openstack/glance-db-sync-5wg47" podUID="03546c3d-2da5-4476-8c22-5a14939cfd80" Nov 24 07:05:09 crc kubenswrapper[4799]: I1124 07:05:09.927936 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-2dz2n-config-bgp5x" podStartSLOduration=9.927912807 podStartE2EDuration="9.927912807s" podCreationTimestamp="2025-11-24 07:05:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:09.921964458 +0000 UTC m=+1055.577946932" watchObservedRunningTime="2025-11-24 07:05:09.927912807 +0000 UTC m=+1055.583895291" Nov 24 07:05:10 crc kubenswrapper[4799]: I1124 07:05:10.246098 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-2dz2n" Nov 24 07:05:10 crc kubenswrapper[4799]: I1124 07:05:10.633444 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 07:05:10 crc kubenswrapper[4799]: I1124 07:05:10.913500 4799 generic.go:334] "Generic (PLEG): container finished" podID="31f1da8c-b203-4427-91cf-08024dbf6f52" containerID="bb57102415efc322a8bcd1bc7aa2a2615715511ee00dafa287e4f8bb115a5c1a" exitCode=0 Nov 24 07:05:10 crc kubenswrapper[4799]: I1124 07:05:10.913596 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n-config-bgp5x" event={"ID":"31f1da8c-b203-4427-91cf-08024dbf6f52","Type":"ContainerDied","Data":"bb57102415efc322a8bcd1bc7aa2a2615715511ee00dafa287e4f8bb115a5c1a"} Nov 24 07:05:10 crc kubenswrapper[4799]: I1124 07:05:10.925883 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"6d52bb110c0d262b0946e49bcf01652ba0241c80b9cad0add83df423d3294499"} Nov 24 07:05:10 crc kubenswrapper[4799]: I1124 07:05:10.925960 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"22c9751adfda0a99f240920e420a888061186c08a987e917be24c8bc2d4dc9d3"} Nov 24 07:05:10 crc kubenswrapper[4799]: I1124 07:05:10.925980 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"8ac4f0ccf8167b086abf61aae9df350c81a1408a802be79619c6747aa2cc47af"} Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.012944 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-v4ct4"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.014350 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v4ct4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.037453 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v4ct4"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.046829 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.116889 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-bbb1-account-create-87l6k"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.118160 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbb1-account-create-87l6k" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.123871 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-v9jnb"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.124237 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6b4c\" (UniqueName: \"kubernetes.io/projected/e00ef351-7e26-4b68-8c9b-a37510eff377-kube-api-access-z6b4c\") pod \"cinder-db-create-v4ct4\" (UID: \"e00ef351-7e26-4b68-8c9b-a37510eff377\") " pod="openstack/cinder-db-create-v4ct4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.124381 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00ef351-7e26-4b68-8c9b-a37510eff377-operator-scripts\") pod \"cinder-db-create-v4ct4\" (UID: \"e00ef351-7e26-4b68-8c9b-a37510eff377\") " pod="openstack/cinder-db-create-v4ct4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.125036 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v9jnb" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.125319 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.139810 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bbb1-account-create-87l6k"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.145582 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-v9jnb"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.229744 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwgrm\" (UniqueName: \"kubernetes.io/projected/e0c34246-d1bb-401c-94e5-084d6064ce5b-kube-api-access-jwgrm\") pod \"barbican-db-create-v9jnb\" (UID: \"e0c34246-d1bb-401c-94e5-084d6064ce5b\") " pod="openstack/barbican-db-create-v9jnb" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.229893 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-operator-scripts\") pod \"cinder-bbb1-account-create-87l6k\" (UID: \"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3\") " pod="openstack/cinder-bbb1-account-create-87l6k" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.230248 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0c34246-d1bb-401c-94e5-084d6064ce5b-operator-scripts\") pod \"barbican-db-create-v9jnb\" (UID: \"e0c34246-d1bb-401c-94e5-084d6064ce5b\") " pod="openstack/barbican-db-create-v9jnb" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.230335 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00ef351-7e26-4b68-8c9b-a37510eff377-operator-scripts\") pod \"cinder-db-create-v4ct4\" (UID: \"e00ef351-7e26-4b68-8c9b-a37510eff377\") " pod="openstack/cinder-db-create-v4ct4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.230550 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6jjs\" (UniqueName: \"kubernetes.io/projected/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-kube-api-access-t6jjs\") pod \"cinder-bbb1-account-create-87l6k\" (UID: \"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3\") " pod="openstack/cinder-bbb1-account-create-87l6k" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.230593 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6b4c\" (UniqueName: \"kubernetes.io/projected/e00ef351-7e26-4b68-8c9b-a37510eff377-kube-api-access-z6b4c\") pod \"cinder-db-create-v4ct4\" (UID: \"e00ef351-7e26-4b68-8c9b-a37510eff377\") " pod="openstack/cinder-db-create-v4ct4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.231353 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00ef351-7e26-4b68-8c9b-a37510eff377-operator-scripts\") pod \"cinder-db-create-v4ct4\" (UID: \"e00ef351-7e26-4b68-8c9b-a37510eff377\") " pod="openstack/cinder-db-create-v4ct4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.258535 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-44f2-account-create-ds5c4"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.263994 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-44f2-account-create-ds5c4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.266697 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.321555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6b4c\" (UniqueName: \"kubernetes.io/projected/e00ef351-7e26-4b68-8c9b-a37510eff377-kube-api-access-z6b4c\") pod \"cinder-db-create-v4ct4\" (UID: \"e00ef351-7e26-4b68-8c9b-a37510eff377\") " pod="openstack/cinder-db-create-v4ct4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.330049 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v4ct4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.333976 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99c90963-48b3-45aa-85cc-f6d51b5702be-operator-scripts\") pod \"barbican-44f2-account-create-ds5c4\" (UID: \"99c90963-48b3-45aa-85cc-f6d51b5702be\") " pod="openstack/barbican-44f2-account-create-ds5c4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.334041 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z8cj\" (UniqueName: \"kubernetes.io/projected/99c90963-48b3-45aa-85cc-f6d51b5702be-kube-api-access-4z8cj\") pod \"barbican-44f2-account-create-ds5c4\" (UID: \"99c90963-48b3-45aa-85cc-f6d51b5702be\") " pod="openstack/barbican-44f2-account-create-ds5c4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.334094 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0c34246-d1bb-401c-94e5-084d6064ce5b-operator-scripts\") pod \"barbican-db-create-v9jnb\" (UID: \"e0c34246-d1bb-401c-94e5-084d6064ce5b\") " pod="openstack/barbican-db-create-v9jnb" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.334200 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6jjs\" (UniqueName: \"kubernetes.io/projected/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-kube-api-access-t6jjs\") pod \"cinder-bbb1-account-create-87l6k\" (UID: \"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3\") " pod="openstack/cinder-bbb1-account-create-87l6k" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.334238 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwgrm\" (UniqueName: \"kubernetes.io/projected/e0c34246-d1bb-401c-94e5-084d6064ce5b-kube-api-access-jwgrm\") pod \"barbican-db-create-v9jnb\" (UID: \"e0c34246-d1bb-401c-94e5-084d6064ce5b\") " pod="openstack/barbican-db-create-v9jnb" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.334280 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-operator-scripts\") pod \"cinder-bbb1-account-create-87l6k\" (UID: \"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3\") " pod="openstack/cinder-bbb1-account-create-87l6k" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.339780 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-44f2-account-create-ds5c4"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.345458 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-operator-scripts\") pod \"cinder-bbb1-account-create-87l6k\" (UID: \"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3\") " pod="openstack/cinder-bbb1-account-create-87l6k" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.348163 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0c34246-d1bb-401c-94e5-084d6064ce5b-operator-scripts\") pod \"barbican-db-create-v9jnb\" (UID: \"e0c34246-d1bb-401c-94e5-084d6064ce5b\") " pod="openstack/barbican-db-create-v9jnb" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.383696 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwgrm\" (UniqueName: \"kubernetes.io/projected/e0c34246-d1bb-401c-94e5-084d6064ce5b-kube-api-access-jwgrm\") pod \"barbican-db-create-v9jnb\" (UID: \"e0c34246-d1bb-401c-94e5-084d6064ce5b\") " pod="openstack/barbican-db-create-v9jnb" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.389267 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6jjs\" (UniqueName: \"kubernetes.io/projected/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-kube-api-access-t6jjs\") pod \"cinder-bbb1-account-create-87l6k\" (UID: \"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3\") " pod="openstack/cinder-bbb1-account-create-87l6k" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.437886 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99c90963-48b3-45aa-85cc-f6d51b5702be-operator-scripts\") pod \"barbican-44f2-account-create-ds5c4\" (UID: \"99c90963-48b3-45aa-85cc-f6d51b5702be\") " pod="openstack/barbican-44f2-account-create-ds5c4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.437947 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z8cj\" (UniqueName: \"kubernetes.io/projected/99c90963-48b3-45aa-85cc-f6d51b5702be-kube-api-access-4z8cj\") pod \"barbican-44f2-account-create-ds5c4\" (UID: \"99c90963-48b3-45aa-85cc-f6d51b5702be\") " pod="openstack/barbican-44f2-account-create-ds5c4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.438873 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99c90963-48b3-45aa-85cc-f6d51b5702be-operator-scripts\") pod \"barbican-44f2-account-create-ds5c4\" (UID: \"99c90963-48b3-45aa-85cc-f6d51b5702be\") " pod="openstack/barbican-44f2-account-create-ds5c4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.450606 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbb1-account-create-87l6k" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.479595 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z8cj\" (UniqueName: \"kubernetes.io/projected/99c90963-48b3-45aa-85cc-f6d51b5702be-kube-api-access-4z8cj\") pod \"barbican-44f2-account-create-ds5c4\" (UID: \"99c90963-48b3-45aa-85cc-f6d51b5702be\") " pod="openstack/barbican-44f2-account-create-ds5c4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.482754 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v9jnb" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.514153 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-hslpd"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.515817 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hslpd" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.550637 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hslpd"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.595940 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2d2f-account-create-mgts2"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.604653 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2d2f-account-create-mgts2" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.606776 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.643322 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2d2f-account-create-mgts2"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.644459 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/207abd4d-7482-4d1e-93b8-abab45c55888-operator-scripts\") pod \"neutron-db-create-hslpd\" (UID: \"207abd4d-7482-4d1e-93b8-abab45c55888\") " pod="openstack/neutron-db-create-hslpd" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.644739 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-44f2-account-create-ds5c4" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.647224 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwsvb\" (UniqueName: \"kubernetes.io/projected/207abd4d-7482-4d1e-93b8-abab45c55888-kube-api-access-lwsvb\") pod \"neutron-db-create-hslpd\" (UID: \"207abd4d-7482-4d1e-93b8-abab45c55888\") " pod="openstack/neutron-db-create-hslpd" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.647655 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-bclrf"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.649801 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.653406 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kggcp" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.653619 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.653743 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.655725 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.688729 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bclrf"] Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.748702 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn4jj\" (UniqueName: \"kubernetes.io/projected/1331194e-76cf-4c49-9d14-294d7b3897bd-kube-api-access-cn4jj\") pod \"neutron-2d2f-account-create-mgts2\" (UID: \"1331194e-76cf-4c49-9d14-294d7b3897bd\") " pod="openstack/neutron-2d2f-account-create-mgts2" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.748747 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/207abd4d-7482-4d1e-93b8-abab45c55888-operator-scripts\") pod \"neutron-db-create-hslpd\" (UID: \"207abd4d-7482-4d1e-93b8-abab45c55888\") " pod="openstack/neutron-db-create-hslpd" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.748769 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-combined-ca-bundle\") pod \"keystone-db-sync-bclrf\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.748786 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwp7v\" (UniqueName: \"kubernetes.io/projected/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-kube-api-access-fwp7v\") pod \"keystone-db-sync-bclrf\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.748808 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1331194e-76cf-4c49-9d14-294d7b3897bd-operator-scripts\") pod \"neutron-2d2f-account-create-mgts2\" (UID: \"1331194e-76cf-4c49-9d14-294d7b3897bd\") " pod="openstack/neutron-2d2f-account-create-mgts2" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.748887 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-config-data\") pod \"keystone-db-sync-bclrf\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.748926 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwsvb\" (UniqueName: \"kubernetes.io/projected/207abd4d-7482-4d1e-93b8-abab45c55888-kube-api-access-lwsvb\") pod \"neutron-db-create-hslpd\" (UID: \"207abd4d-7482-4d1e-93b8-abab45c55888\") " pod="openstack/neutron-db-create-hslpd" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.751182 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/207abd4d-7482-4d1e-93b8-abab45c55888-operator-scripts\") pod \"neutron-db-create-hslpd\" (UID: \"207abd4d-7482-4d1e-93b8-abab45c55888\") " pod="openstack/neutron-db-create-hslpd" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.775201 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwsvb\" (UniqueName: \"kubernetes.io/projected/207abd4d-7482-4d1e-93b8-abab45c55888-kube-api-access-lwsvb\") pod \"neutron-db-create-hslpd\" (UID: \"207abd4d-7482-4d1e-93b8-abab45c55888\") " pod="openstack/neutron-db-create-hslpd" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.853129 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-config-data\") pod \"keystone-db-sync-bclrf\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.853835 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn4jj\" (UniqueName: \"kubernetes.io/projected/1331194e-76cf-4c49-9d14-294d7b3897bd-kube-api-access-cn4jj\") pod \"neutron-2d2f-account-create-mgts2\" (UID: \"1331194e-76cf-4c49-9d14-294d7b3897bd\") " pod="openstack/neutron-2d2f-account-create-mgts2" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.853881 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-combined-ca-bundle\") pod \"keystone-db-sync-bclrf\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.853912 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwp7v\" (UniqueName: \"kubernetes.io/projected/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-kube-api-access-fwp7v\") pod \"keystone-db-sync-bclrf\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.853939 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1331194e-76cf-4c49-9d14-294d7b3897bd-operator-scripts\") pod \"neutron-2d2f-account-create-mgts2\" (UID: \"1331194e-76cf-4c49-9d14-294d7b3897bd\") " pod="openstack/neutron-2d2f-account-create-mgts2" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.856204 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1331194e-76cf-4c49-9d14-294d7b3897bd-operator-scripts\") pod \"neutron-2d2f-account-create-mgts2\" (UID: \"1331194e-76cf-4c49-9d14-294d7b3897bd\") " pod="openstack/neutron-2d2f-account-create-mgts2" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.862588 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-combined-ca-bundle\") pod \"keystone-db-sync-bclrf\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.862701 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-config-data\") pod \"keystone-db-sync-bclrf\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.874704 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn4jj\" (UniqueName: \"kubernetes.io/projected/1331194e-76cf-4c49-9d14-294d7b3897bd-kube-api-access-cn4jj\") pod \"neutron-2d2f-account-create-mgts2\" (UID: \"1331194e-76cf-4c49-9d14-294d7b3897bd\") " pod="openstack/neutron-2d2f-account-create-mgts2" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.882451 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwp7v\" (UniqueName: \"kubernetes.io/projected/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-kube-api-access-fwp7v\") pod \"keystone-db-sync-bclrf\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.901691 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hslpd" Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.958647 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"977a35126c8112db8b3dddf63c96480f5494a7696b5982d214d42310a46c7b83"} Nov 24 07:05:11 crc kubenswrapper[4799]: I1124 07:05:11.958682 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"479a19c5d4397919f77e002bce98bc8cfe8b88d861166ad64dbe89337e0e2126"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.011414 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2d2f-account-create-mgts2" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.027787 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.038249 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v4ct4"] Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.166938 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bbb1-account-create-87l6k"] Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.310098 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-v9jnb"] Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.461899 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-44f2-account-create-ds5c4"] Nov 24 07:05:12 crc kubenswrapper[4799]: E1124 07:05:12.466929 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"container-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"container-replicator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75\\\"\", failed to \"StartContainer\" for \"container-auditor\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75\\\"\", failed to \"StartContainer\" for \"container-updater\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75\\\"\"]" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.665712 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.680087 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hslpd"] Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.694836 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2d2f-account-create-mgts2"] Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.729886 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bclrf"] Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.773542 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "31f1da8c-b203-4427-91cf-08024dbf6f52" (UID: "31f1da8c-b203-4427-91cf-08024dbf6f52"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.773926 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run-ovn\") pod \"31f1da8c-b203-4427-91cf-08024dbf6f52\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.776398 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-log-ovn\") pod \"31f1da8c-b203-4427-91cf-08024dbf6f52\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.776504 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-additional-scripts\") pod \"31f1da8c-b203-4427-91cf-08024dbf6f52\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.776607 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-scripts\") pod \"31f1da8c-b203-4427-91cf-08024dbf6f52\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.776716 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttbt8\" (UniqueName: \"kubernetes.io/projected/31f1da8c-b203-4427-91cf-08024dbf6f52-kube-api-access-ttbt8\") pod \"31f1da8c-b203-4427-91cf-08024dbf6f52\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.776883 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run\") pod \"31f1da8c-b203-4427-91cf-08024dbf6f52\" (UID: \"31f1da8c-b203-4427-91cf-08024dbf6f52\") " Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.776716 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "31f1da8c-b203-4427-91cf-08024dbf6f52" (UID: "31f1da8c-b203-4427-91cf-08024dbf6f52"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.777342 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run" (OuterVolumeSpecName: "var-run") pod "31f1da8c-b203-4427-91cf-08024dbf6f52" (UID: "31f1da8c-b203-4427-91cf-08024dbf6f52"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.778346 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-scripts" (OuterVolumeSpecName: "scripts") pod "31f1da8c-b203-4427-91cf-08024dbf6f52" (UID: "31f1da8c-b203-4427-91cf-08024dbf6f52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.778751 4799 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.779681 4799 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.779790 4799 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/31f1da8c-b203-4427-91cf-08024dbf6f52-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.784061 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "31f1da8c-b203-4427-91cf-08024dbf6f52" (UID: "31f1da8c-b203-4427-91cf-08024dbf6f52"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.816904 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31f1da8c-b203-4427-91cf-08024dbf6f52-kube-api-access-ttbt8" (OuterVolumeSpecName: "kube-api-access-ttbt8") pod "31f1da8c-b203-4427-91cf-08024dbf6f52" (UID: "31f1da8c-b203-4427-91cf-08024dbf6f52"). InnerVolumeSpecName "kube-api-access-ttbt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.882503 4799 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.883752 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31f1da8c-b203-4427-91cf-08024dbf6f52-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.883858 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttbt8\" (UniqueName: \"kubernetes.io/projected/31f1da8c-b203-4427-91cf-08024dbf6f52-kube-api-access-ttbt8\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.968321 4799 generic.go:334] "Generic (PLEG): container finished" podID="e00ef351-7e26-4b68-8c9b-a37510eff377" containerID="3a79a583f1fe1d2c0c47fb1685562489608c16c24f320f11e04b10758e7e52b0" exitCode=0 Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.968433 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v4ct4" event={"ID":"e00ef351-7e26-4b68-8c9b-a37510eff377","Type":"ContainerDied","Data":"3a79a583f1fe1d2c0c47fb1685562489608c16c24f320f11e04b10758e7e52b0"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.968463 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v4ct4" event={"ID":"e00ef351-7e26-4b68-8c9b-a37510eff377","Type":"ContainerStarted","Data":"126aca9a4fec3da8d2302cd96b39bf5c4b76fb7fe64f3434d4870ed751b2782c"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.970280 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bclrf" event={"ID":"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100","Type":"ContainerStarted","Data":"fc4858e0dfb33b463c5bb68291eeee099559922da08ecc54e838a197ebd71c43"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.974563 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bbb1-account-create-87l6k" event={"ID":"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3","Type":"ContainerStarted","Data":"cdd7465ba59afe1601f7bbb0d441c7457c2da76592769f351a34b40d11ac5f9a"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.974610 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bbb1-account-create-87l6k" event={"ID":"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3","Type":"ContainerStarted","Data":"b684ea5e1d7ff64d9cfbf0923871a16b0aeb8006ac9173965cde699b08167231"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.989320 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"57ac5adb075544c22ae8aa32fb5357e1f72bc9a58d99ba17c8bb148b66dade15"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.990826 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"815111e7fbfc05e4da9663fc01168d0881dc15d3f9b8fe7ec1ce5a3f38cc6d40"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.994920 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2d2f-account-create-mgts2" event={"ID":"1331194e-76cf-4c49-9d14-294d7b3897bd","Type":"ContainerStarted","Data":"6200c3568ecb120918ba43f8e8a656d9d1acb4db6b319da813f226c1af7e3025"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.995039 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2d2f-account-create-mgts2" event={"ID":"1331194e-76cf-4c49-9d14-294d7b3897bd","Type":"ContainerStarted","Data":"20ff450af7de36c25e28b1baa033fe1ec97fcb26e572eae8682034d82391f82e"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.996156 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-44f2-account-create-ds5c4" event={"ID":"99c90963-48b3-45aa-85cc-f6d51b5702be","Type":"ContainerStarted","Data":"adcebf3d8165db5c2482742ffe61dd12640c95e10c666abdf160cac3dd0d6b08"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.996259 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-44f2-account-create-ds5c4" event={"ID":"99c90963-48b3-45aa-85cc-f6d51b5702be","Type":"ContainerStarted","Data":"4ccae23b05d7ac2ad2f29141ff5eddebc953efc78cd1ca082ed216f6bb857213"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.999363 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n-config-bgp5x" event={"ID":"31f1da8c-b203-4427-91cf-08024dbf6f52","Type":"ContainerDied","Data":"3d105a895e4f10cac21184e257510d6b1183cc0c0f542a1b1b935ca293f7d8bf"} Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.999401 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d105a895e4f10cac21184e257510d6b1183cc0c0f542a1b1b935ca293f7d8bf" Nov 24 07:05:12 crc kubenswrapper[4799]: I1124 07:05:12.999462 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n-config-bgp5x" Nov 24 07:05:13 crc kubenswrapper[4799]: E1124 07:05:13.000910 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"container-server\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75\\\"\", failed to \"StartContainer\" for \"container-replicator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75\\\"\", failed to \"StartContainer\" for \"container-auditor\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75\\\"\", failed to \"StartContainer\" for \"container-updater\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75\\\"\"]" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.001523 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hslpd" event={"ID":"207abd4d-7482-4d1e-93b8-abab45c55888","Type":"ContainerStarted","Data":"7a67d14359667b49d385e2378597910f70c85b133bae02d6c75678b330114be2"} Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.001606 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hslpd" event={"ID":"207abd4d-7482-4d1e-93b8-abab45c55888","Type":"ContainerStarted","Data":"22393346e23053534b941eae3e365ce93776c35d210fec67928957a7d57bc716"} Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.006129 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v9jnb" event={"ID":"e0c34246-d1bb-401c-94e5-084d6064ce5b","Type":"ContainerStarted","Data":"a21f45844786ca81306aa2947d5cce7e325c670a6b991864ca7472ab29d8c0b9"} Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.006193 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v9jnb" event={"ID":"e0c34246-d1bb-401c-94e5-084d6064ce5b","Type":"ContainerStarted","Data":"c452743cbfde40ea058e0045e9cfa1fdba2a281ab30d793f7861467e2b83c817"} Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.019912 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-bbb1-account-create-87l6k" podStartSLOduration=2.019889936 podStartE2EDuration="2.019889936s" podCreationTimestamp="2025-11-24 07:05:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:13.015645796 +0000 UTC m=+1058.671628270" watchObservedRunningTime="2025-11-24 07:05:13.019889936 +0000 UTC m=+1058.675872410" Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.044436 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-v9jnb" podStartSLOduration=2.044417249 podStartE2EDuration="2.044417249s" podCreationTimestamp="2025-11-24 07:05:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:13.033919262 +0000 UTC m=+1058.689901726" watchObservedRunningTime="2025-11-24 07:05:13.044417249 +0000 UTC m=+1058.700399723" Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.063932 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-44f2-account-create-ds5c4" podStartSLOduration=2.063911651 podStartE2EDuration="2.063911651s" podCreationTimestamp="2025-11-24 07:05:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:13.057501549 +0000 UTC m=+1058.713484023" watchObservedRunningTime="2025-11-24 07:05:13.063911651 +0000 UTC m=+1058.719894115" Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.095034 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-hslpd" podStartSLOduration=2.09500944 podStartE2EDuration="2.09500944s" podCreationTimestamp="2025-11-24 07:05:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:13.079974155 +0000 UTC m=+1058.735956629" watchObservedRunningTime="2025-11-24 07:05:13.09500944 +0000 UTC m=+1058.750991914" Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.107719 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-2d2f-account-create-mgts2" podStartSLOduration=2.107700759 podStartE2EDuration="2.107700759s" podCreationTimestamp="2025-11-24 07:05:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:13.101804763 +0000 UTC m=+1058.757787237" watchObservedRunningTime="2025-11-24 07:05:13.107700759 +0000 UTC m=+1058.763683233" Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.839792 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-2dz2n-config-bgp5x"] Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.848089 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-2dz2n-config-bgp5x"] Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.927251 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2dz2n-config-bjtqp"] Nov 24 07:05:13 crc kubenswrapper[4799]: E1124 07:05:13.927738 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31f1da8c-b203-4427-91cf-08024dbf6f52" containerName="ovn-config" Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.927758 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="31f1da8c-b203-4427-91cf-08024dbf6f52" containerName="ovn-config" Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.928038 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="31f1da8c-b203-4427-91cf-08024dbf6f52" containerName="ovn-config" Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.928785 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.935008 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 07:05:13 crc kubenswrapper[4799]: I1124 07:05:13.945017 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2dz2n-config-bjtqp"] Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.005649 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.005704 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-log-ovn\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.005750 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-additional-scripts\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.005789 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run-ovn\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.005832 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-scripts\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.005865 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5rps\" (UniqueName: \"kubernetes.io/projected/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-kube-api-access-n5rps\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.024127 4799 generic.go:334] "Generic (PLEG): container finished" podID="207abd4d-7482-4d1e-93b8-abab45c55888" containerID="7a67d14359667b49d385e2378597910f70c85b133bae02d6c75678b330114be2" exitCode=0 Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.024223 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hslpd" event={"ID":"207abd4d-7482-4d1e-93b8-abab45c55888","Type":"ContainerDied","Data":"7a67d14359667b49d385e2378597910f70c85b133bae02d6c75678b330114be2"} Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.029451 4799 generic.go:334] "Generic (PLEG): container finished" podID="e0c34246-d1bb-401c-94e5-084d6064ce5b" containerID="a21f45844786ca81306aa2947d5cce7e325c670a6b991864ca7472ab29d8c0b9" exitCode=0 Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.029609 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v9jnb" event={"ID":"e0c34246-d1bb-401c-94e5-084d6064ce5b","Type":"ContainerDied","Data":"a21f45844786ca81306aa2947d5cce7e325c670a6b991864ca7472ab29d8c0b9"} Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.033716 4799 generic.go:334] "Generic (PLEG): container finished" podID="a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3" containerID="cdd7465ba59afe1601f7bbb0d441c7457c2da76592769f351a34b40d11ac5f9a" exitCode=0 Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.033950 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bbb1-account-create-87l6k" event={"ID":"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3","Type":"ContainerDied","Data":"cdd7465ba59afe1601f7bbb0d441c7457c2da76592769f351a34b40d11ac5f9a"} Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.040603 4799 generic.go:334] "Generic (PLEG): container finished" podID="1331194e-76cf-4c49-9d14-294d7b3897bd" containerID="6200c3568ecb120918ba43f8e8a656d9d1acb4db6b319da813f226c1af7e3025" exitCode=0 Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.040695 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2d2f-account-create-mgts2" event={"ID":"1331194e-76cf-4c49-9d14-294d7b3897bd","Type":"ContainerDied","Data":"6200c3568ecb120918ba43f8e8a656d9d1acb4db6b319da813f226c1af7e3025"} Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.043643 4799 generic.go:334] "Generic (PLEG): container finished" podID="99c90963-48b3-45aa-85cc-f6d51b5702be" containerID="adcebf3d8165db5c2482742ffe61dd12640c95e10c666abdf160cac3dd0d6b08" exitCode=0 Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.043752 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-44f2-account-create-ds5c4" event={"ID":"99c90963-48b3-45aa-85cc-f6d51b5702be","Type":"ContainerDied","Data":"adcebf3d8165db5c2482742ffe61dd12640c95e10c666abdf160cac3dd0d6b08"} Nov 24 07:05:14 crc kubenswrapper[4799]: E1124 07:05:14.048549 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"container-server\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75\\\"\", failed to \"StartContainer\" for \"container-replicator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75\\\"\", failed to \"StartContainer\" for \"container-auditor\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75\\\"\", failed to \"StartContainer\" for \"container-updater\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75\\\"\"]" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.109454 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-additional-scripts\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.109571 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run-ovn\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.109624 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-scripts\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.109677 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5rps\" (UniqueName: \"kubernetes.io/projected/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-kube-api-access-n5rps\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.109816 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.109887 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-log-ovn\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.110408 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-log-ovn\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.110720 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run-ovn\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.110932 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.111286 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-additional-scripts\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.112903 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-scripts\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.135692 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5rps\" (UniqueName: \"kubernetes.io/projected/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-kube-api-access-n5rps\") pod \"ovn-controller-2dz2n-config-bjtqp\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.253197 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.514993 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v4ct4" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.626072 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00ef351-7e26-4b68-8c9b-a37510eff377-operator-scripts\") pod \"e00ef351-7e26-4b68-8c9b-a37510eff377\" (UID: \"e00ef351-7e26-4b68-8c9b-a37510eff377\") " Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.626118 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6b4c\" (UniqueName: \"kubernetes.io/projected/e00ef351-7e26-4b68-8c9b-a37510eff377-kube-api-access-z6b4c\") pod \"e00ef351-7e26-4b68-8c9b-a37510eff377\" (UID: \"e00ef351-7e26-4b68-8c9b-a37510eff377\") " Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.627422 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e00ef351-7e26-4b68-8c9b-a37510eff377-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e00ef351-7e26-4b68-8c9b-a37510eff377" (UID: "e00ef351-7e26-4b68-8c9b-a37510eff377"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.638375 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e00ef351-7e26-4b68-8c9b-a37510eff377-kube-api-access-z6b4c" (OuterVolumeSpecName: "kube-api-access-z6b4c") pod "e00ef351-7e26-4b68-8c9b-a37510eff377" (UID: "e00ef351-7e26-4b68-8c9b-a37510eff377"). InnerVolumeSpecName "kube-api-access-z6b4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.672027 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2dz2n-config-bjtqp"] Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.728521 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00ef351-7e26-4b68-8c9b-a37510eff377-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:14 crc kubenswrapper[4799]: I1124 07:05:14.728557 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6b4c\" (UniqueName: \"kubernetes.io/projected/e00ef351-7e26-4b68-8c9b-a37510eff377-kube-api-access-z6b4c\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.054566 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v4ct4" event={"ID":"e00ef351-7e26-4b68-8c9b-a37510eff377","Type":"ContainerDied","Data":"126aca9a4fec3da8d2302cd96b39bf5c4b76fb7fe64f3434d4870ed751b2782c"} Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.054605 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="126aca9a4fec3da8d2302cd96b39bf5c4b76fb7fe64f3434d4870ed751b2782c" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.054630 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v4ct4" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.057598 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n-config-bjtqp" event={"ID":"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f","Type":"ContainerStarted","Data":"2440fe858d8868b1c4904a513e11ec3d1af8a9378f9ca998eb5b15957accc132"} Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.057627 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n-config-bjtqp" event={"ID":"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f","Type":"ContainerStarted","Data":"e5e391faf78e4394d47b6512e9f62990ce9d42884cc5efe7afe0c1c99ee7bd0d"} Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.081955 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-2dz2n-config-bjtqp" podStartSLOduration=2.081936813 podStartE2EDuration="2.081936813s" podCreationTimestamp="2025-11-24 07:05:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:15.081620444 +0000 UTC m=+1060.737602928" watchObservedRunningTime="2025-11-24 07:05:15.081936813 +0000 UTC m=+1060.737919287" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.402485 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2d2f-account-create-mgts2" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.446940 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn4jj\" (UniqueName: \"kubernetes.io/projected/1331194e-76cf-4c49-9d14-294d7b3897bd-kube-api-access-cn4jj\") pod \"1331194e-76cf-4c49-9d14-294d7b3897bd\" (UID: \"1331194e-76cf-4c49-9d14-294d7b3897bd\") " Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.447114 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1331194e-76cf-4c49-9d14-294d7b3897bd-operator-scripts\") pod \"1331194e-76cf-4c49-9d14-294d7b3897bd\" (UID: \"1331194e-76cf-4c49-9d14-294d7b3897bd\") " Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.448752 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1331194e-76cf-4c49-9d14-294d7b3897bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1331194e-76cf-4c49-9d14-294d7b3897bd" (UID: "1331194e-76cf-4c49-9d14-294d7b3897bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.464305 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1331194e-76cf-4c49-9d14-294d7b3897bd-kube-api-access-cn4jj" (OuterVolumeSpecName: "kube-api-access-cn4jj") pod "1331194e-76cf-4c49-9d14-294d7b3897bd" (UID: "1331194e-76cf-4c49-9d14-294d7b3897bd"). InnerVolumeSpecName "kube-api-access-cn4jj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.554252 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn4jj\" (UniqueName: \"kubernetes.io/projected/1331194e-76cf-4c49-9d14-294d7b3897bd-kube-api-access-cn4jj\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.554336 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1331194e-76cf-4c49-9d14-294d7b3897bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.590106 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hslpd" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.590960 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbb1-account-create-87l6k" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.600677 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-44f2-account-create-ds5c4" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.608421 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v9jnb" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.654565 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31f1da8c-b203-4427-91cf-08024dbf6f52" path="/var/lib/kubelet/pods/31f1da8c-b203-4427-91cf-08024dbf6f52/volumes" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.655143 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0c34246-d1bb-401c-94e5-084d6064ce5b-operator-scripts\") pod \"e0c34246-d1bb-401c-94e5-084d6064ce5b\" (UID: \"e0c34246-d1bb-401c-94e5-084d6064ce5b\") " Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.655232 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z8cj\" (UniqueName: \"kubernetes.io/projected/99c90963-48b3-45aa-85cc-f6d51b5702be-kube-api-access-4z8cj\") pod \"99c90963-48b3-45aa-85cc-f6d51b5702be\" (UID: \"99c90963-48b3-45aa-85cc-f6d51b5702be\") " Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.655307 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99c90963-48b3-45aa-85cc-f6d51b5702be-operator-scripts\") pod \"99c90963-48b3-45aa-85cc-f6d51b5702be\" (UID: \"99c90963-48b3-45aa-85cc-f6d51b5702be\") " Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.655380 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwsvb\" (UniqueName: \"kubernetes.io/projected/207abd4d-7482-4d1e-93b8-abab45c55888-kube-api-access-lwsvb\") pod \"207abd4d-7482-4d1e-93b8-abab45c55888\" (UID: \"207abd4d-7482-4d1e-93b8-abab45c55888\") " Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.655417 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6jjs\" (UniqueName: \"kubernetes.io/projected/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-kube-api-access-t6jjs\") pod \"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3\" (UID: \"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3\") " Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.655471 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwgrm\" (UniqueName: \"kubernetes.io/projected/e0c34246-d1bb-401c-94e5-084d6064ce5b-kube-api-access-jwgrm\") pod \"e0c34246-d1bb-401c-94e5-084d6064ce5b\" (UID: \"e0c34246-d1bb-401c-94e5-084d6064ce5b\") " Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.655571 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-operator-scripts\") pod \"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3\" (UID: \"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3\") " Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.655616 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/207abd4d-7482-4d1e-93b8-abab45c55888-operator-scripts\") pod \"207abd4d-7482-4d1e-93b8-abab45c55888\" (UID: \"207abd4d-7482-4d1e-93b8-abab45c55888\") " Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.658871 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/207abd4d-7482-4d1e-93b8-abab45c55888-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "207abd4d-7482-4d1e-93b8-abab45c55888" (UID: "207abd4d-7482-4d1e-93b8-abab45c55888"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.661781 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0c34246-d1bb-401c-94e5-084d6064ce5b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e0c34246-d1bb-401c-94e5-084d6064ce5b" (UID: "e0c34246-d1bb-401c-94e5-084d6064ce5b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.662186 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c90963-48b3-45aa-85cc-f6d51b5702be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99c90963-48b3-45aa-85cc-f6d51b5702be" (UID: "99c90963-48b3-45aa-85cc-f6d51b5702be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.662328 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3" (UID: "a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.668217 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-kube-api-access-t6jjs" (OuterVolumeSpecName: "kube-api-access-t6jjs") pod "a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3" (UID: "a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3"). InnerVolumeSpecName "kube-api-access-t6jjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.669450 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0c34246-d1bb-401c-94e5-084d6064ce5b-kube-api-access-jwgrm" (OuterVolumeSpecName: "kube-api-access-jwgrm") pod "e0c34246-d1bb-401c-94e5-084d6064ce5b" (UID: "e0c34246-d1bb-401c-94e5-084d6064ce5b"). InnerVolumeSpecName "kube-api-access-jwgrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.678362 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/207abd4d-7482-4d1e-93b8-abab45c55888-kube-api-access-lwsvb" (OuterVolumeSpecName: "kube-api-access-lwsvb") pod "207abd4d-7482-4d1e-93b8-abab45c55888" (UID: "207abd4d-7482-4d1e-93b8-abab45c55888"). InnerVolumeSpecName "kube-api-access-lwsvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.695250 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99c90963-48b3-45aa-85cc-f6d51b5702be-kube-api-access-4z8cj" (OuterVolumeSpecName: "kube-api-access-4z8cj") pod "99c90963-48b3-45aa-85cc-f6d51b5702be" (UID: "99c90963-48b3-45aa-85cc-f6d51b5702be"). InnerVolumeSpecName "kube-api-access-4z8cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.758454 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0c34246-d1bb-401c-94e5-084d6064ce5b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.758507 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z8cj\" (UniqueName: \"kubernetes.io/projected/99c90963-48b3-45aa-85cc-f6d51b5702be-kube-api-access-4z8cj\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.758528 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99c90963-48b3-45aa-85cc-f6d51b5702be-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.758541 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwsvb\" (UniqueName: \"kubernetes.io/projected/207abd4d-7482-4d1e-93b8-abab45c55888-kube-api-access-lwsvb\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.758553 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6jjs\" (UniqueName: \"kubernetes.io/projected/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-kube-api-access-t6jjs\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.758571 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwgrm\" (UniqueName: \"kubernetes.io/projected/e0c34246-d1bb-401c-94e5-084d6064ce5b-kube-api-access-jwgrm\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.758583 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:15 crc kubenswrapper[4799]: I1124 07:05:15.758594 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/207abd4d-7482-4d1e-93b8-abab45c55888-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.107865 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbb1-account-create-87l6k" Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.108063 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bbb1-account-create-87l6k" event={"ID":"a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3","Type":"ContainerDied","Data":"b684ea5e1d7ff64d9cfbf0923871a16b0aeb8006ac9173965cde699b08167231"} Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.108129 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b684ea5e1d7ff64d9cfbf0923871a16b0aeb8006ac9173965cde699b08167231" Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.111579 4799 generic.go:334] "Generic (PLEG): container finished" podID="f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f" containerID="2440fe858d8868b1c4904a513e11ec3d1af8a9378f9ca998eb5b15957accc132" exitCode=0 Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.112059 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n-config-bjtqp" event={"ID":"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f","Type":"ContainerDied","Data":"2440fe858d8868b1c4904a513e11ec3d1af8a9378f9ca998eb5b15957accc132"} Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.114909 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2d2f-account-create-mgts2" event={"ID":"1331194e-76cf-4c49-9d14-294d7b3897bd","Type":"ContainerDied","Data":"20ff450af7de36c25e28b1baa033fe1ec97fcb26e572eae8682034d82391f82e"} Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.115247 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20ff450af7de36c25e28b1baa033fe1ec97fcb26e572eae8682034d82391f82e" Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.115139 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2d2f-account-create-mgts2" Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.117383 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-44f2-account-create-ds5c4" Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.117376 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-44f2-account-create-ds5c4" event={"ID":"99c90963-48b3-45aa-85cc-f6d51b5702be","Type":"ContainerDied","Data":"4ccae23b05d7ac2ad2f29141ff5eddebc953efc78cd1ca082ed216f6bb857213"} Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.117549 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ccae23b05d7ac2ad2f29141ff5eddebc953efc78cd1ca082ed216f6bb857213" Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.120654 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hslpd" event={"ID":"207abd4d-7482-4d1e-93b8-abab45c55888","Type":"ContainerDied","Data":"22393346e23053534b941eae3e365ce93776c35d210fec67928957a7d57bc716"} Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.120687 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22393346e23053534b941eae3e365ce93776c35d210fec67928957a7d57bc716" Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.120732 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hslpd" Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.125508 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v9jnb" event={"ID":"e0c34246-d1bb-401c-94e5-084d6064ce5b","Type":"ContainerDied","Data":"c452743cbfde40ea058e0045e9cfa1fdba2a281ab30d793f7861467e2b83c817"} Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.125541 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c452743cbfde40ea058e0045e9cfa1fdba2a281ab30d793f7861467e2b83c817" Nov 24 07:05:16 crc kubenswrapper[4799]: I1124 07:05:16.125587 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v9jnb" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.678049 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.758540 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run\") pod \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.758607 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-additional-scripts\") pod \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.758710 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-scripts\") pod \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.758884 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run-ovn\") pod \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.758921 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-log-ovn\") pod \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.758971 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5rps\" (UniqueName: \"kubernetes.io/projected/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-kube-api-access-n5rps\") pod \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\" (UID: \"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f\") " Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.758991 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run" (OuterVolumeSpecName: "var-run") pod "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f" (UID: "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.759208 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f" (UID: "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.759237 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f" (UID: "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.760014 4799 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.760046 4799 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.760062 4799 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.760182 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f" (UID: "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.761158 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-scripts" (OuterVolumeSpecName: "scripts") pod "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f" (UID: "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.764295 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-kube-api-access-n5rps" (OuterVolumeSpecName: "kube-api-access-n5rps") pod "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f" (UID: "f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f"). InnerVolumeSpecName "kube-api-access-n5rps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.862189 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5rps\" (UniqueName: \"kubernetes.io/projected/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-kube-api-access-n5rps\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.862233 4799 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:18 crc kubenswrapper[4799]: I1124 07:05:18.862248 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:19 crc kubenswrapper[4799]: I1124 07:05:19.157975 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bclrf" event={"ID":"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100","Type":"ContainerStarted","Data":"6048bb73c3893b671ede7f9dc617529b1d172e7ca9f7861936ac9ca8e5189813"} Nov 24 07:05:19 crc kubenswrapper[4799]: I1124 07:05:19.162621 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n-config-bjtqp" event={"ID":"f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f","Type":"ContainerDied","Data":"e5e391faf78e4394d47b6512e9f62990ce9d42884cc5efe7afe0c1c99ee7bd0d"} Nov 24 07:05:19 crc kubenswrapper[4799]: I1124 07:05:19.162682 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5e391faf78e4394d47b6512e9f62990ce9d42884cc5efe7afe0c1c99ee7bd0d" Nov 24 07:05:19 crc kubenswrapper[4799]: I1124 07:05:19.162685 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n-config-bjtqp" Nov 24 07:05:19 crc kubenswrapper[4799]: I1124 07:05:19.196223 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-bclrf" podStartSLOduration=2.468300759 podStartE2EDuration="8.196192798s" podCreationTimestamp="2025-11-24 07:05:11 +0000 UTC" firstStartedPulling="2025-11-24 07:05:12.784821276 +0000 UTC m=+1058.440803750" lastFinishedPulling="2025-11-24 07:05:18.512713315 +0000 UTC m=+1064.168695789" observedRunningTime="2025-11-24 07:05:19.182403078 +0000 UTC m=+1064.838385592" watchObservedRunningTime="2025-11-24 07:05:19.196192798 +0000 UTC m=+1064.852175282" Nov 24 07:05:19 crc kubenswrapper[4799]: I1124 07:05:19.770297 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-2dz2n-config-bjtqp"] Nov 24 07:05:19 crc kubenswrapper[4799]: I1124 07:05:19.781677 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-2dz2n-config-bjtqp"] Nov 24 07:05:21 crc kubenswrapper[4799]: I1124 07:05:21.639076 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f" path="/var/lib/kubelet/pods/f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f/volumes" Nov 24 07:05:22 crc kubenswrapper[4799]: I1124 07:05:22.199957 4799 generic.go:334] "Generic (PLEG): container finished" podID="e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100" containerID="6048bb73c3893b671ede7f9dc617529b1d172e7ca9f7861936ac9ca8e5189813" exitCode=0 Nov 24 07:05:22 crc kubenswrapper[4799]: I1124 07:05:22.200027 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bclrf" event={"ID":"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100","Type":"ContainerDied","Data":"6048bb73c3893b671ede7f9dc617529b1d172e7ca9f7861936ac9ca8e5189813"} Nov 24 07:05:23 crc kubenswrapper[4799]: I1124 07:05:23.591552 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:23 crc kubenswrapper[4799]: I1124 07:05:23.659674 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-combined-ca-bundle\") pod \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " Nov 24 07:05:23 crc kubenswrapper[4799]: I1124 07:05:23.659775 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-config-data\") pod \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " Nov 24 07:05:23 crc kubenswrapper[4799]: I1124 07:05:23.659825 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwp7v\" (UniqueName: \"kubernetes.io/projected/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-kube-api-access-fwp7v\") pod \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\" (UID: \"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100\") " Nov 24 07:05:23 crc kubenswrapper[4799]: I1124 07:05:23.667487 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-kube-api-access-fwp7v" (OuterVolumeSpecName: "kube-api-access-fwp7v") pod "e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100" (UID: "e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100"). InnerVolumeSpecName "kube-api-access-fwp7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:23 crc kubenswrapper[4799]: I1124 07:05:23.695796 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100" (UID: "e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:23 crc kubenswrapper[4799]: I1124 07:05:23.728639 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-config-data" (OuterVolumeSpecName: "config-data") pod "e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100" (UID: "e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:23 crc kubenswrapper[4799]: I1124 07:05:23.761501 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:23 crc kubenswrapper[4799]: I1124 07:05:23.761530 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:23 crc kubenswrapper[4799]: I1124 07:05:23.761541 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwp7v\" (UniqueName: \"kubernetes.io/projected/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100-kube-api-access-fwp7v\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.227865 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bclrf" event={"ID":"e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100","Type":"ContainerDied","Data":"fc4858e0dfb33b463c5bb68291eeee099559922da08ecc54e838a197ebd71c43"} Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.228121 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc4858e0dfb33b463c5bb68291eeee099559922da08ecc54e838a197ebd71c43" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.228175 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bclrf" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489125 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c9f7b7b67-x99d9"] Nov 24 07:05:24 crc kubenswrapper[4799]: E1124 07:05:24.489528 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100" containerName="keystone-db-sync" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489548 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100" containerName="keystone-db-sync" Nov 24 07:05:24 crc kubenswrapper[4799]: E1124 07:05:24.489571 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1331194e-76cf-4c49-9d14-294d7b3897bd" containerName="mariadb-account-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489578 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1331194e-76cf-4c49-9d14-294d7b3897bd" containerName="mariadb-account-create" Nov 24 07:05:24 crc kubenswrapper[4799]: E1124 07:05:24.489589 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3" containerName="mariadb-account-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489595 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3" containerName="mariadb-account-create" Nov 24 07:05:24 crc kubenswrapper[4799]: E1124 07:05:24.489609 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207abd4d-7482-4d1e-93b8-abab45c55888" containerName="mariadb-database-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489615 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="207abd4d-7482-4d1e-93b8-abab45c55888" containerName="mariadb-database-create" Nov 24 07:05:24 crc kubenswrapper[4799]: E1124 07:05:24.489624 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e00ef351-7e26-4b68-8c9b-a37510eff377" containerName="mariadb-database-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489630 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e00ef351-7e26-4b68-8c9b-a37510eff377" containerName="mariadb-database-create" Nov 24 07:05:24 crc kubenswrapper[4799]: E1124 07:05:24.489646 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f" containerName="ovn-config" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489651 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f" containerName="ovn-config" Nov 24 07:05:24 crc kubenswrapper[4799]: E1124 07:05:24.489662 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99c90963-48b3-45aa-85cc-f6d51b5702be" containerName="mariadb-account-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489668 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="99c90963-48b3-45aa-85cc-f6d51b5702be" containerName="mariadb-account-create" Nov 24 07:05:24 crc kubenswrapper[4799]: E1124 07:05:24.489678 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0c34246-d1bb-401c-94e5-084d6064ce5b" containerName="mariadb-database-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489685 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0c34246-d1bb-401c-94e5-084d6064ce5b" containerName="mariadb-database-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489835 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e00ef351-7e26-4b68-8c9b-a37510eff377" containerName="mariadb-database-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489861 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="207abd4d-7482-4d1e-93b8-abab45c55888" containerName="mariadb-database-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489872 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1331194e-76cf-4c49-9d14-294d7b3897bd" containerName="mariadb-account-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489882 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100" containerName="keystone-db-sync" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489891 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0c34246-d1bb-401c-94e5-084d6064ce5b" containerName="mariadb-database-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489898 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3" containerName="mariadb-account-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489909 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9e3832e-4db1-4bdf-a3a7-6a823d1ce22f" containerName="ovn-config" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.489921 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="99c90963-48b3-45aa-85cc-f6d51b5702be" containerName="mariadb-account-create" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.490882 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.497029 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-627r7"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.498835 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.505277 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.505475 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.505571 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.505894 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.508345 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kggcp" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.512453 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-627r7"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.534742 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c9f7b7b67-x99d9"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.581293 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjksc\" (UniqueName: \"kubernetes.io/projected/c841e6ef-485e-4534-9f55-95daab50f2cf-kube-api-access-jjksc\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.581366 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-sb\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.581402 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-scripts\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.581423 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-nb\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.581448 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-dns-svc\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.581465 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-config\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.581481 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95vpk\" (UniqueName: \"kubernetes.io/projected/26acd2c4-7349-431a-9732-738924445a38-kube-api-access-95vpk\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.581495 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-credential-keys\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.581512 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-fernet-keys\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.581534 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-combined-ca-bundle\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.581550 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-config-data\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.667335 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-qfbzk"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.668449 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.676059 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-tsmnv" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.676390 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.676632 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.683165 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-scripts\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.683436 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-nb\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.683529 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-dns-svc\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.683607 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95vpk\" (UniqueName: \"kubernetes.io/projected/26acd2c4-7349-431a-9732-738924445a38-kube-api-access-95vpk\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.683688 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-config\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.683764 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-credential-keys\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.683858 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-fernet-keys\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.683996 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-combined-ca-bundle\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.684078 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-config-data\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.684238 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjksc\" (UniqueName: \"kubernetes.io/projected/c841e6ef-485e-4534-9f55-95daab50f2cf-kube-api-access-jjksc\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.684354 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-sb\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.685284 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-sb\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.685458 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-config\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.689570 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-nb\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.689600 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-dns-svc\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.696809 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-fernet-keys\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.702887 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-scripts\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.703099 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-credential-keys\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.706884 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-qfbzk"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.719138 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjksc\" (UniqueName: \"kubernetes.io/projected/c841e6ef-485e-4534-9f55-95daab50f2cf-kube-api-access-jjksc\") pod \"dnsmasq-dns-c9f7b7b67-x99d9\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.720034 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-combined-ca-bundle\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.720101 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-kjxxs"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.721488 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.723966 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-config-data\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.750351 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.750738 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rsrhk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.750754 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kjxxs"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.750906 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.762177 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95vpk\" (UniqueName: \"kubernetes.io/projected/26acd2c4-7349-431a-9732-738924445a38-kube-api-access-95vpk\") pod \"keystone-bootstrap-627r7\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.776895 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.779218 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.786127 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.786407 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.788273 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a987addd-6874-4499-9ea1-27ec674bf28a-etc-machine-id\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.788306 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dwj2\" (UniqueName: \"kubernetes.io/projected/a987addd-6874-4499-9ea1-27ec674bf28a-kube-api-access-4dwj2\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.788367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-scripts\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.788400 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-config-data\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.788432 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-db-sync-config-data\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.788486 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-combined-ca-bundle\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.791707 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.820998 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-gpvs7"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.821967 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.830986 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.831806 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d296j" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.846652 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-gpvs7"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.891220 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.891769 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.892885 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-config-data\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.892912 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-scripts\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.892953 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-scripts\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.892971 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-run-httpd\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.892998 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-config-data\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.893038 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.893061 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjj5s\" (UniqueName: \"kubernetes.io/projected/71fffc24-2640-46ca-92aa-f5a2926f5864-kube-api-access-zjj5s\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.893100 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-db-sync-config-data\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.893130 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-combined-ca-bundle\") pod \"neutron-db-sync-kjxxs\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.893149 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-log-httpd\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.893197 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-combined-ca-bundle\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.893217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-config\") pod \"neutron-db-sync-kjxxs\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.893237 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.893273 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkdr2\" (UniqueName: \"kubernetes.io/projected/b3cd7765-7004-4a36-8331-029e35639568-kube-api-access-tkdr2\") pod \"neutron-db-sync-kjxxs\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.893310 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a987addd-6874-4499-9ea1-27ec674bf28a-etc-machine-id\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.893343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dwj2\" (UniqueName: \"kubernetes.io/projected/a987addd-6874-4499-9ea1-27ec674bf28a-kube-api-access-4dwj2\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.918378 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c9f7b7b67-x99d9"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.919593 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a987addd-6874-4499-9ea1-27ec674bf28a-etc-machine-id\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.922543 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dwj2\" (UniqueName: \"kubernetes.io/projected/a987addd-6874-4499-9ea1-27ec674bf28a-kube-api-access-4dwj2\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.923280 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-combined-ca-bundle\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.931765 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-db-sync-config-data\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.932127 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-scripts\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.949901 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-config-data\") pod \"cinder-db-sync-qfbzk\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.979293 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-794df4974f-r8fcp"] Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.980697 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.994893 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-combined-ca-bundle\") pod \"barbican-db-sync-gpvs7\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.994984 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-config-data\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.995011 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-scripts\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.995031 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-run-httpd\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.995066 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.995091 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqx9j\" (UniqueName: \"kubernetes.io/projected/57aab8d2-548c-47cf-9beb-bbe3958ebce6-kube-api-access-vqx9j\") pod \"barbican-db-sync-gpvs7\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.995119 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjj5s\" (UniqueName: \"kubernetes.io/projected/71fffc24-2640-46ca-92aa-f5a2926f5864-kube-api-access-zjj5s\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.995177 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-combined-ca-bundle\") pod \"neutron-db-sync-kjxxs\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.995205 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-log-httpd\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.995244 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-config\") pod \"neutron-db-sync-kjxxs\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.995266 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-db-sync-config-data\") pod \"barbican-db-sync-gpvs7\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.995290 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:24 crc kubenswrapper[4799]: I1124 07:05:24.995312 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkdr2\" (UniqueName: \"kubernetes.io/projected/b3cd7765-7004-4a36-8331-029e35639568-kube-api-access-tkdr2\") pod \"neutron-db-sync-kjxxs\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.001738 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-run-httpd\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.002371 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-log-httpd\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.003712 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-config-data\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.023484 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-794df4974f-r8fcp"] Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.025410 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.025410 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.025520 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-config\") pod \"neutron-db-sync-kjxxs\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.025997 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-combined-ca-bundle\") pod \"neutron-db-sync-kjxxs\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.034496 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkdr2\" (UniqueName: \"kubernetes.io/projected/b3cd7765-7004-4a36-8331-029e35639568-kube-api-access-tkdr2\") pod \"neutron-db-sync-kjxxs\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.043458 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-scripts\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.044444 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjj5s\" (UniqueName: \"kubernetes.io/projected/71fffc24-2640-46ca-92aa-f5a2926f5864-kube-api-access-zjj5s\") pod \"ceilometer-0\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " pod="openstack/ceilometer-0" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.099240 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.100288 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-combined-ca-bundle\") pod \"barbican-db-sync-gpvs7\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.100337 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjxrc\" (UniqueName: \"kubernetes.io/projected/048ad289-b5a0-44cf-936f-11680a5dadd9-kube-api-access-hjxrc\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.100372 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqx9j\" (UniqueName: \"kubernetes.io/projected/57aab8d2-548c-47cf-9beb-bbe3958ebce6-kube-api-access-vqx9j\") pod \"barbican-db-sync-gpvs7\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.100392 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-dns-svc\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.100408 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-nb\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.100444 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-config\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.100476 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-sb\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.100491 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-db-sync-config-data\") pod \"barbican-db-sync-gpvs7\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.117458 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-db-sync-config-data\") pod \"barbican-db-sync-gpvs7\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.124936 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-combined-ca-bundle\") pod \"barbican-db-sync-gpvs7\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.144203 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.163314 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqx9j\" (UniqueName: \"kubernetes.io/projected/57aab8d2-548c-47cf-9beb-bbe3958ebce6-kube-api-access-vqx9j\") pod \"barbican-db-sync-gpvs7\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.163617 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-hj8ls"] Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.168984 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.174270 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.180728 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.181938 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hj8ls"] Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.189347 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.189577 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.203747 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xtntc" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.211819 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-sb\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.211947 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjxrc\" (UniqueName: \"kubernetes.io/projected/048ad289-b5a0-44cf-936f-11680a5dadd9-kube-api-access-hjxrc\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.211979 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-dns-svc\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.212021 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-nb\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.212062 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-config\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.213048 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-config\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.213611 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-dns-svc\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.213769 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-sb\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.214091 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-nb\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.259673 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjxrc\" (UniqueName: \"kubernetes.io/projected/048ad289-b5a0-44cf-936f-11680a5dadd9-kube-api-access-hjxrc\") pod \"dnsmasq-dns-794df4974f-r8fcp\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.313757 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-scripts\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.313829 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-combined-ca-bundle\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.313861 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ccb00cb-56e5-4588-8c60-88b52e92721e-logs\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.313896 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z52f7\" (UniqueName: \"kubernetes.io/projected/0ccb00cb-56e5-4588-8c60-88b52e92721e-kube-api-access-z52f7\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.313971 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-config-data\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.336751 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.406617 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5wg47" event={"ID":"03546c3d-2da5-4476-8c22-5a14939cfd80","Type":"ContainerStarted","Data":"f9312779e4995c99fe48adfc65b9f134a79e29c78cf5915c12e2453b6490a6cd"} Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.416785 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-config-data\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.416911 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-scripts\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.416989 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-combined-ca-bundle\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.417029 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ccb00cb-56e5-4588-8c60-88b52e92721e-logs\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.417118 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z52f7\" (UniqueName: \"kubernetes.io/projected/0ccb00cb-56e5-4588-8c60-88b52e92721e-kube-api-access-z52f7\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.422411 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ccb00cb-56e5-4588-8c60-88b52e92721e-logs\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.437173 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-scripts\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.441909 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-combined-ca-bundle\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.441966 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-config-data\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.491534 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z52f7\" (UniqueName: \"kubernetes.io/projected/0ccb00cb-56e5-4588-8c60-88b52e92721e-kube-api-access-z52f7\") pod \"placement-db-sync-hj8ls\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.507582 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-5wg47" podStartSLOduration=2.539360192 podStartE2EDuration="36.507559239s" podCreationTimestamp="2025-11-24 07:04:49 +0000 UTC" firstStartedPulling="2025-11-24 07:04:50.170032269 +0000 UTC m=+1035.826014743" lastFinishedPulling="2025-11-24 07:05:24.138231316 +0000 UTC m=+1069.794213790" observedRunningTime="2025-11-24 07:05:25.446877063 +0000 UTC m=+1071.102859537" watchObservedRunningTime="2025-11-24 07:05:25.507559239 +0000 UTC m=+1071.163541713" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.554759 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:25 crc kubenswrapper[4799]: I1124 07:05:25.783154 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c9f7b7b67-x99d9"] Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.040033 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-627r7"] Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.356555 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-794df4974f-r8fcp"] Nov 24 07:05:26 crc kubenswrapper[4799]: W1124 07:05:26.362088 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod048ad289_b5a0_44cf_936f_11680a5dadd9.slice/crio-97a9a1c89aba5c1915e9d195206b7c54914aa923ce02c88b4b0ad4307f8a9d69 WatchSource:0}: Error finding container 97a9a1c89aba5c1915e9d195206b7c54914aa923ce02c88b4b0ad4307f8a9d69: Status 404 returned error can't find the container with id 97a9a1c89aba5c1915e9d195206b7c54914aa923ce02c88b4b0ad4307f8a9d69 Nov 24 07:05:26 crc kubenswrapper[4799]: W1124 07:05:26.366683 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda987addd_6874_4499_9ea1_27ec674bf28a.slice/crio-6e5ca720cf0da438c6b4e13370ddf0b7b58035de88d9a5a1b82ac52f8c4966c5 WatchSource:0}: Error finding container 6e5ca720cf0da438c6b4e13370ddf0b7b58035de88d9a5a1b82ac52f8c4966c5: Status 404 returned error can't find the container with id 6e5ca720cf0da438c6b4e13370ddf0b7b58035de88d9a5a1b82ac52f8c4966c5 Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.370488 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-qfbzk"] Nov 24 07:05:26 crc kubenswrapper[4799]: W1124 07:05:26.378600 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57aab8d2_548c_47cf_9beb_bbe3958ebce6.slice/crio-b4e6d1c789d5ad1997ff8598a7fddd4efe69b58699f9faa41cebffae7648a034 WatchSource:0}: Error finding container b4e6d1c789d5ad1997ff8598a7fddd4efe69b58699f9faa41cebffae7648a034: Status 404 returned error can't find the container with id b4e6d1c789d5ad1997ff8598a7fddd4efe69b58699f9faa41cebffae7648a034 Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.396769 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-gpvs7"] Nov 24 07:05:26 crc kubenswrapper[4799]: W1124 07:05:26.404797 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3cd7765_7004_4a36_8331_029e35639568.slice/crio-73f84e58d77d5f3e78ee96d30231e9e72378cb9a4a6b3421b1ce97a433141013 WatchSource:0}: Error finding container 73f84e58d77d5f3e78ee96d30231e9e72378cb9a4a6b3421b1ce97a433141013: Status 404 returned error can't find the container with id 73f84e58d77d5f3e78ee96d30231e9e72378cb9a4a6b3421b1ce97a433141013 Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.409920 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.420630 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kjxxs"] Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.427052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kjxxs" event={"ID":"b3cd7765-7004-4a36-8331-029e35639568","Type":"ContainerStarted","Data":"73f84e58d77d5f3e78ee96d30231e9e72378cb9a4a6b3421b1ce97a433141013"} Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.429768 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gpvs7" event={"ID":"57aab8d2-548c-47cf-9beb-bbe3958ebce6","Type":"ContainerStarted","Data":"b4e6d1c789d5ad1997ff8598a7fddd4efe69b58699f9faa41cebffae7648a034"} Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.431625 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qfbzk" event={"ID":"a987addd-6874-4499-9ea1-27ec674bf28a","Type":"ContainerStarted","Data":"6e5ca720cf0da438c6b4e13370ddf0b7b58035de88d9a5a1b82ac52f8c4966c5"} Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.434078 4799 generic.go:334] "Generic (PLEG): container finished" podID="c841e6ef-485e-4534-9f55-95daab50f2cf" containerID="860846f622d8979dd655a22b57751ac5db8172e2d45ed1e5274730889de922e1" exitCode=0 Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.434140 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" event={"ID":"c841e6ef-485e-4534-9f55-95daab50f2cf","Type":"ContainerDied","Data":"860846f622d8979dd655a22b57751ac5db8172e2d45ed1e5274730889de922e1"} Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.434161 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" event={"ID":"c841e6ef-485e-4534-9f55-95daab50f2cf","Type":"ContainerStarted","Data":"51eee41906ededed02ea3033529e19a9834456ab71c56d878c0a4749b663819d"} Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.439356 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" event={"ID":"048ad289-b5a0-44cf-936f-11680a5dadd9","Type":"ContainerStarted","Data":"97a9a1c89aba5c1915e9d195206b7c54914aa923ce02c88b4b0ad4307f8a9d69"} Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.458578 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hj8ls"] Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.476658 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-627r7" event={"ID":"26acd2c4-7349-431a-9732-738924445a38","Type":"ContainerStarted","Data":"175f63c46417051e3c6f78ba7dd59612a708ca3b330b37ba4144be88999e43af"} Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.476705 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-627r7" event={"ID":"26acd2c4-7349-431a-9732-738924445a38","Type":"ContainerStarted","Data":"0f058b0fac7e5c0ad629311f112333b673bc528fe451ee33d9149cbe7af3bac3"} Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.495400 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-627r7" podStartSLOduration=2.495377961 podStartE2EDuration="2.495377961s" podCreationTimestamp="2025-11-24 07:05:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:26.491553813 +0000 UTC m=+1072.147536287" watchObservedRunningTime="2025-11-24 07:05:26.495377961 +0000 UTC m=+1072.151360435" Nov 24 07:05:26 crc kubenswrapper[4799]: I1124 07:05:26.939669 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.017237 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.073584 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjksc\" (UniqueName: \"kubernetes.io/projected/c841e6ef-485e-4534-9f55-95daab50f2cf-kube-api-access-jjksc\") pod \"c841e6ef-485e-4534-9f55-95daab50f2cf\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.073694 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-nb\") pod \"c841e6ef-485e-4534-9f55-95daab50f2cf\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.073806 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-dns-svc\") pod \"c841e6ef-485e-4534-9f55-95daab50f2cf\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.073930 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-sb\") pod \"c841e6ef-485e-4534-9f55-95daab50f2cf\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.073980 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-config\") pod \"c841e6ef-485e-4534-9f55-95daab50f2cf\" (UID: \"c841e6ef-485e-4534-9f55-95daab50f2cf\") " Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.093258 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c841e6ef-485e-4534-9f55-95daab50f2cf-kube-api-access-jjksc" (OuterVolumeSpecName: "kube-api-access-jjksc") pod "c841e6ef-485e-4534-9f55-95daab50f2cf" (UID: "c841e6ef-485e-4534-9f55-95daab50f2cf"). InnerVolumeSpecName "kube-api-access-jjksc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.115727 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-config" (OuterVolumeSpecName: "config") pod "c841e6ef-485e-4534-9f55-95daab50f2cf" (UID: "c841e6ef-485e-4534-9f55-95daab50f2cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.118263 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c841e6ef-485e-4534-9f55-95daab50f2cf" (UID: "c841e6ef-485e-4534-9f55-95daab50f2cf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.119857 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c841e6ef-485e-4534-9f55-95daab50f2cf" (UID: "c841e6ef-485e-4534-9f55-95daab50f2cf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.131455 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c841e6ef-485e-4534-9f55-95daab50f2cf" (UID: "c841e6ef-485e-4534-9f55-95daab50f2cf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.176157 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjksc\" (UniqueName: \"kubernetes.io/projected/c841e6ef-485e-4534-9f55-95daab50f2cf-kube-api-access-jjksc\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.176209 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.176224 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.176236 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.176247 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c841e6ef-485e-4534-9f55-95daab50f2cf-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.488937 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.488833 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9f7b7b67-x99d9" event={"ID":"c841e6ef-485e-4534-9f55-95daab50f2cf","Type":"ContainerDied","Data":"51eee41906ededed02ea3033529e19a9834456ab71c56d878c0a4749b663819d"} Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.489049 4799 scope.go:117] "RemoveContainer" containerID="860846f622d8979dd655a22b57751ac5db8172e2d45ed1e5274730889de922e1" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.493193 4799 generic.go:334] "Generic (PLEG): container finished" podID="048ad289-b5a0-44cf-936f-11680a5dadd9" containerID="35128b4273c0252968ccf6ddba0596ddf3f6fef887658802bbfe8f22c6c9b8c8" exitCode=0 Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.493242 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" event={"ID":"048ad289-b5a0-44cf-936f-11680a5dadd9","Type":"ContainerDied","Data":"35128b4273c0252968ccf6ddba0596ddf3f6fef887658802bbfe8f22c6c9b8c8"} Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.500129 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hj8ls" event={"ID":"0ccb00cb-56e5-4588-8c60-88b52e92721e","Type":"ContainerStarted","Data":"ba24fd7e0bdaa34c711156e9f57931de002441ec120ede0972f000b6f0099667"} Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.511238 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kjxxs" event={"ID":"b3cd7765-7004-4a36-8331-029e35639568","Type":"ContainerStarted","Data":"2f8c8bdc7ce403fb12cce9fdf8de873b21760d472a20b313828bee7d09f25e41"} Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.535059 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71fffc24-2640-46ca-92aa-f5a2926f5864","Type":"ContainerStarted","Data":"b0ed93876a566b5169ba2512b5b3cc1b3c5678feba1004428c91b4b189a351ed"} Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.556626 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-kjxxs" podStartSLOduration=3.556608868 podStartE2EDuration="3.556608868s" podCreationTimestamp="2025-11-24 07:05:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:27.556114124 +0000 UTC m=+1073.212096598" watchObservedRunningTime="2025-11-24 07:05:27.556608868 +0000 UTC m=+1073.212591342" Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.693947 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c9f7b7b67-x99d9"] Nov 24 07:05:27 crc kubenswrapper[4799]: I1124 07:05:27.714145 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c9f7b7b67-x99d9"] Nov 24 07:05:28 crc kubenswrapper[4799]: I1124 07:05:28.575056 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" event={"ID":"048ad289-b5a0-44cf-936f-11680a5dadd9","Type":"ContainerStarted","Data":"2cc6db81131f6ab67f4802806a3f8c76bc87db675b5dcd87a04ea0d21d57a7ea"} Nov 24 07:05:28 crc kubenswrapper[4799]: I1124 07:05:28.575222 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:28 crc kubenswrapper[4799]: I1124 07:05:28.613294 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" podStartSLOduration=4.613270757 podStartE2EDuration="4.613270757s" podCreationTimestamp="2025-11-24 07:05:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:28.601987798 +0000 UTC m=+1074.257970272" watchObservedRunningTime="2025-11-24 07:05:28.613270757 +0000 UTC m=+1074.269253231" Nov 24 07:05:29 crc kubenswrapper[4799]: I1124 07:05:29.599952 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"9b90579a000b03e6467dabd651cbce2abb14ef949a00cfdd47d7596c85645587"} Nov 24 07:05:29 crc kubenswrapper[4799]: I1124 07:05:29.600354 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"c01da2c24a6c25ef981a4bf81e1ef3766e00137d2cf3e37995a2696e9e28633c"} Nov 24 07:05:29 crc kubenswrapper[4799]: I1124 07:05:29.600372 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"5e46da699973a5e22e83b9aaac7e0e2ff7525ee80f32ff63ebc1fc46a7e36544"} Nov 24 07:05:29 crc kubenswrapper[4799]: I1124 07:05:29.648283 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c841e6ef-485e-4534-9f55-95daab50f2cf" path="/var/lib/kubelet/pods/c841e6ef-485e-4534-9f55-95daab50f2cf/volumes" Nov 24 07:05:30 crc kubenswrapper[4799]: I1124 07:05:30.640942 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerStarted","Data":"c8e47fdf47fa1c6b2fdcb1d77c84b445e782cfd00343ab8b1883a72bed9feebd"} Nov 24 07:05:30 crc kubenswrapper[4799]: I1124 07:05:30.648111 4799 generic.go:334] "Generic (PLEG): container finished" podID="26acd2c4-7349-431a-9732-738924445a38" containerID="175f63c46417051e3c6f78ba7dd59612a708ca3b330b37ba4144be88999e43af" exitCode=0 Nov 24 07:05:30 crc kubenswrapper[4799]: I1124 07:05:30.648157 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-627r7" event={"ID":"26acd2c4-7349-431a-9732-738924445a38","Type":"ContainerDied","Data":"175f63c46417051e3c6f78ba7dd59612a708ca3b330b37ba4144be88999e43af"} Nov 24 07:05:30 crc kubenswrapper[4799]: I1124 07:05:30.701140 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.64190183 podStartE2EDuration="54.701119913s" podCreationTimestamp="2025-11-24 07:04:36 +0000 UTC" firstStartedPulling="2025-11-24 07:04:54.246195136 +0000 UTC m=+1039.902177610" lastFinishedPulling="2025-11-24 07:05:28.305413219 +0000 UTC m=+1073.961395693" observedRunningTime="2025-11-24 07:05:30.697607383 +0000 UTC m=+1076.353589857" watchObservedRunningTime="2025-11-24 07:05:30.701119913 +0000 UTC m=+1076.357102387" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.042557 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794df4974f-r8fcp"] Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.042766 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" podUID="048ad289-b5a0-44cf-936f-11680a5dadd9" containerName="dnsmasq-dns" containerID="cri-o://2cc6db81131f6ab67f4802806a3f8c76bc87db675b5dcd87a04ea0d21d57a7ea" gracePeriod=10 Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.081147 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bdb874957-w6lwg"] Nov 24 07:05:31 crc kubenswrapper[4799]: E1124 07:05:31.081643 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c841e6ef-485e-4534-9f55-95daab50f2cf" containerName="init" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.081664 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c841e6ef-485e-4534-9f55-95daab50f2cf" containerName="init" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.081902 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c841e6ef-485e-4534-9f55-95daab50f2cf" containerName="init" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.083007 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.087552 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.101799 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bdb874957-w6lwg"] Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.178773 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-svc\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.179014 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-nb\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.179071 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-config\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.179094 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpk4q\" (UniqueName: \"kubernetes.io/projected/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-kube-api-access-jpk4q\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.179128 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-swift-storage-0\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.179167 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-sb\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.280334 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-swift-storage-0\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.280404 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-sb\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.280526 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-svc\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.280581 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-nb\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.280679 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-config\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.280700 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpk4q\" (UniqueName: \"kubernetes.io/projected/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-kube-api-access-jpk4q\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.281500 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-swift-storage-0\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.281548 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-sb\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.282142 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-nb\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.282393 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-svc\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.282734 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-config\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.316524 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpk4q\" (UniqueName: \"kubernetes.io/projected/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-kube-api-access-jpk4q\") pod \"dnsmasq-dns-6bdb874957-w6lwg\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.458330 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.669246 4799 generic.go:334] "Generic (PLEG): container finished" podID="048ad289-b5a0-44cf-936f-11680a5dadd9" containerID="2cc6db81131f6ab67f4802806a3f8c76bc87db675b5dcd87a04ea0d21d57a7ea" exitCode=0 Nov 24 07:05:31 crc kubenswrapper[4799]: I1124 07:05:31.669285 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" event={"ID":"048ad289-b5a0-44cf-936f-11680a5dadd9","Type":"ContainerDied","Data":"2cc6db81131f6ab67f4802806a3f8c76bc87db675b5dcd87a04ea0d21d57a7ea"} Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.235947 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.354188 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-fernet-keys\") pod \"26acd2c4-7349-431a-9732-738924445a38\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.354295 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-credential-keys\") pod \"26acd2c4-7349-431a-9732-738924445a38\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.354379 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-combined-ca-bundle\") pod \"26acd2c4-7349-431a-9732-738924445a38\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.354546 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-scripts\") pod \"26acd2c4-7349-431a-9732-738924445a38\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.354596 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-config-data\") pod \"26acd2c4-7349-431a-9732-738924445a38\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.354652 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95vpk\" (UniqueName: \"kubernetes.io/projected/26acd2c4-7349-431a-9732-738924445a38-kube-api-access-95vpk\") pod \"26acd2c4-7349-431a-9732-738924445a38\" (UID: \"26acd2c4-7349-431a-9732-738924445a38\") " Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.362321 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "26acd2c4-7349-431a-9732-738924445a38" (UID: "26acd2c4-7349-431a-9732-738924445a38"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.363215 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "26acd2c4-7349-431a-9732-738924445a38" (UID: "26acd2c4-7349-431a-9732-738924445a38"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.368841 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26acd2c4-7349-431a-9732-738924445a38-kube-api-access-95vpk" (OuterVolumeSpecName: "kube-api-access-95vpk") pod "26acd2c4-7349-431a-9732-738924445a38" (UID: "26acd2c4-7349-431a-9732-738924445a38"). InnerVolumeSpecName "kube-api-access-95vpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.378315 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-scripts" (OuterVolumeSpecName: "scripts") pod "26acd2c4-7349-431a-9732-738924445a38" (UID: "26acd2c4-7349-431a-9732-738924445a38"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.397068 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26acd2c4-7349-431a-9732-738924445a38" (UID: "26acd2c4-7349-431a-9732-738924445a38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.458109 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-config-data" (OuterVolumeSpecName: "config-data") pod "26acd2c4-7349-431a-9732-738924445a38" (UID: "26acd2c4-7349-431a-9732-738924445a38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.459750 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.459829 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.459877 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95vpk\" (UniqueName: \"kubernetes.io/projected/26acd2c4-7349-431a-9732-738924445a38-kube-api-access-95vpk\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.459892 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.459901 4799 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.459911 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26acd2c4-7349-431a-9732-738924445a38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.705686 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-627r7" event={"ID":"26acd2c4-7349-431a-9732-738924445a38","Type":"ContainerDied","Data":"0f058b0fac7e5c0ad629311f112333b673bc528fe451ee33d9149cbe7af3bac3"} Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.705743 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f058b0fac7e5c0ad629311f112333b673bc528fe451ee33d9149cbe7af3bac3" Nov 24 07:05:34 crc kubenswrapper[4799]: I1124 07:05:34.705816 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-627r7" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.434170 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-627r7"] Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.441380 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-627r7"] Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.524477 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-h5vcf"] Nov 24 07:05:35 crc kubenswrapper[4799]: E1124 07:05:35.525152 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26acd2c4-7349-431a-9732-738924445a38" containerName="keystone-bootstrap" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.525175 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="26acd2c4-7349-431a-9732-738924445a38" containerName="keystone-bootstrap" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.525408 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="26acd2c4-7349-431a-9732-738924445a38" containerName="keystone-bootstrap" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.526201 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.528753 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.529252 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.531332 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.531753 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.533524 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kggcp" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.542156 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-h5vcf"] Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.651770 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26acd2c4-7349-431a-9732-738924445a38" path="/var/lib/kubelet/pods/26acd2c4-7349-431a-9732-738924445a38/volumes" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.684109 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-combined-ca-bundle\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.684162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjls4\" (UniqueName: \"kubernetes.io/projected/fe55fcc2-0636-4524-a2f5-8127ab456570-kube-api-access-zjls4\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.684251 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-fernet-keys\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.684287 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-credential-keys\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.684359 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-scripts\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.684378 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-config-data\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.785938 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-combined-ca-bundle\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.785991 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjls4\" (UniqueName: \"kubernetes.io/projected/fe55fcc2-0636-4524-a2f5-8127ab456570-kube-api-access-zjls4\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.786065 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-fernet-keys\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.786102 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-credential-keys\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.786144 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-scripts\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.786160 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-config-data\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.793185 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-config-data\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.793644 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-scripts\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.797549 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-fernet-keys\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.805835 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-combined-ca-bundle\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.806095 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-credential-keys\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.815325 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjls4\" (UniqueName: \"kubernetes.io/projected/fe55fcc2-0636-4524-a2f5-8127ab456570-kube-api-access-zjls4\") pod \"keystone-bootstrap-h5vcf\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:35 crc kubenswrapper[4799]: I1124 07:05:35.857782 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:40 crc kubenswrapper[4799]: I1124 07:05:40.338446 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" podUID="048ad289-b5a0-44cf-936f-11680a5dadd9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: i/o timeout" Nov 24 07:05:42 crc kubenswrapper[4799]: I1124 07:05:42.787596 4799 generic.go:334] "Generic (PLEG): container finished" podID="03546c3d-2da5-4476-8c22-5a14939cfd80" containerID="f9312779e4995c99fe48adfc65b9f134a79e29c78cf5915c12e2453b6490a6cd" exitCode=0 Nov 24 07:05:42 crc kubenswrapper[4799]: I1124 07:05:42.787710 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5wg47" event={"ID":"03546c3d-2da5-4476-8c22-5a14939cfd80","Type":"ContainerDied","Data":"f9312779e4995c99fe48adfc65b9f134a79e29c78cf5915c12e2453b6490a6cd"} Nov 24 07:05:45 crc kubenswrapper[4799]: I1124 07:05:45.339314 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" podUID="048ad289-b5a0-44cf-936f-11680a5dadd9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: i/o timeout" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.433302 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.439570 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5wg47" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.541122 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-nb\") pod \"048ad289-b5a0-44cf-936f-11680a5dadd9\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.541206 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-dns-svc\") pod \"048ad289-b5a0-44cf-936f-11680a5dadd9\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.541228 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-sb\") pod \"048ad289-b5a0-44cf-936f-11680a5dadd9\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.541288 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjxrc\" (UniqueName: \"kubernetes.io/projected/048ad289-b5a0-44cf-936f-11680a5dadd9-kube-api-access-hjxrc\") pod \"048ad289-b5a0-44cf-936f-11680a5dadd9\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.541403 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgdjm\" (UniqueName: \"kubernetes.io/projected/03546c3d-2da5-4476-8c22-5a14939cfd80-kube-api-access-bgdjm\") pod \"03546c3d-2da5-4476-8c22-5a14939cfd80\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.541431 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-combined-ca-bundle\") pod \"03546c3d-2da5-4476-8c22-5a14939cfd80\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.541475 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-config-data\") pod \"03546c3d-2da5-4476-8c22-5a14939cfd80\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.541503 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-db-sync-config-data\") pod \"03546c3d-2da5-4476-8c22-5a14939cfd80\" (UID: \"03546c3d-2da5-4476-8c22-5a14939cfd80\") " Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.541538 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-config\") pod \"048ad289-b5a0-44cf-936f-11680a5dadd9\" (UID: \"048ad289-b5a0-44cf-936f-11680a5dadd9\") " Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.549328 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "03546c3d-2da5-4476-8c22-5a14939cfd80" (UID: "03546c3d-2da5-4476-8c22-5a14939cfd80"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.549373 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03546c3d-2da5-4476-8c22-5a14939cfd80-kube-api-access-bgdjm" (OuterVolumeSpecName: "kube-api-access-bgdjm") pod "03546c3d-2da5-4476-8c22-5a14939cfd80" (UID: "03546c3d-2da5-4476-8c22-5a14939cfd80"). InnerVolumeSpecName "kube-api-access-bgdjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.563076 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048ad289-b5a0-44cf-936f-11680a5dadd9-kube-api-access-hjxrc" (OuterVolumeSpecName: "kube-api-access-hjxrc") pod "048ad289-b5a0-44cf-936f-11680a5dadd9" (UID: "048ad289-b5a0-44cf-936f-11680a5dadd9"). InnerVolumeSpecName "kube-api-access-hjxrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.589802 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-config" (OuterVolumeSpecName: "config") pod "048ad289-b5a0-44cf-936f-11680a5dadd9" (UID: "048ad289-b5a0-44cf-936f-11680a5dadd9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.591979 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03546c3d-2da5-4476-8c22-5a14939cfd80" (UID: "03546c3d-2da5-4476-8c22-5a14939cfd80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.596245 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "048ad289-b5a0-44cf-936f-11680a5dadd9" (UID: "048ad289-b5a0-44cf-936f-11680a5dadd9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.598399 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "048ad289-b5a0-44cf-936f-11680a5dadd9" (UID: "048ad289-b5a0-44cf-936f-11680a5dadd9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.601228 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "048ad289-b5a0-44cf-936f-11680a5dadd9" (UID: "048ad289-b5a0-44cf-936f-11680a5dadd9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.607426 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-config-data" (OuterVolumeSpecName: "config-data") pod "03546c3d-2da5-4476-8c22-5a14939cfd80" (UID: "03546c3d-2da5-4476-8c22-5a14939cfd80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.652643 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgdjm\" (UniqueName: \"kubernetes.io/projected/03546c3d-2da5-4476-8c22-5a14939cfd80-kube-api-access-bgdjm\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.652711 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.652726 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.652739 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/03546c3d-2da5-4476-8c22-5a14939cfd80-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.652751 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.652764 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.652777 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.652788 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048ad289-b5a0-44cf-936f-11680a5dadd9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.652799 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjxrc\" (UniqueName: \"kubernetes.io/projected/048ad289-b5a0-44cf-936f-11680a5dadd9-kube-api-access-hjxrc\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.835328 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5wg47" event={"ID":"03546c3d-2da5-4476-8c22-5a14939cfd80","Type":"ContainerDied","Data":"f57d8fd5133b831ae3570505d1e8e07f7c1a52dbf6e18b8f97414bb752636596"} Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.835372 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5wg47" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.835381 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f57d8fd5133b831ae3570505d1e8e07f7c1a52dbf6e18b8f97414bb752636596" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.837575 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" event={"ID":"048ad289-b5a0-44cf-936f-11680a5dadd9","Type":"ContainerDied","Data":"97a9a1c89aba5c1915e9d195206b7c54914aa923ce02c88b4b0ad4307f8a9d69"} Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.837604 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.838202 4799 scope.go:117] "RemoveContainer" containerID="2cc6db81131f6ab67f4802806a3f8c76bc87db675b5dcd87a04ea0d21d57a7ea" Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.869625 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794df4974f-r8fcp"] Nov 24 07:05:47 crc kubenswrapper[4799]: I1124 07:05:47.879988 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-794df4974f-r8fcp"] Nov 24 07:05:48 crc kubenswrapper[4799]: I1124 07:05:48.440731 4799 scope.go:117] "RemoveContainer" containerID="35128b4273c0252968ccf6ddba0596ddf3f6fef887658802bbfe8f22c6c9b8c8" Nov 24 07:05:48 crc kubenswrapper[4799]: E1124 07:05:48.458530 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879" Nov 24 07:05:48 crc kubenswrapper[4799]: E1124 07:05:48.459048 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4dwj2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-qfbzk_openstack(a987addd-6874-4499-9ea1-27ec674bf28a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 07:05:48 crc kubenswrapper[4799]: E1124 07:05:48.460621 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-qfbzk" podUID="a987addd-6874-4499-9ea1-27ec674bf28a" Nov 24 07:05:48 crc kubenswrapper[4799]: I1124 07:05:48.866271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hj8ls" event={"ID":"0ccb00cb-56e5-4588-8c60-88b52e92721e","Type":"ContainerStarted","Data":"ae5034ddd416e340224878800975aaca1a7d4c87688395fae504d9fb486b6005"} Nov 24 07:05:48 crc kubenswrapper[4799]: I1124 07:05:48.879662 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gpvs7" event={"ID":"57aab8d2-548c-47cf-9beb-bbe3958ebce6","Type":"ContainerStarted","Data":"34b20cfd09f348a8c1c28edcf64d822d884d0adade4ab63b86fe2ea2d6edf26d"} Nov 24 07:05:48 crc kubenswrapper[4799]: I1124 07:05:48.883394 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bdb874957-w6lwg"] Nov 24 07:05:48 crc kubenswrapper[4799]: I1124 07:05:48.886233 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71fffc24-2640-46ca-92aa-f5a2926f5864","Type":"ContainerStarted","Data":"bd3eb6eaeec25a6e9fe8c5bb6789ed9b5917e54bffc90e07f048592536b9c6c4"} Nov 24 07:05:48 crc kubenswrapper[4799]: E1124 07:05:48.894616 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879\\\"\"" pod="openstack/cinder-db-sync-qfbzk" podUID="a987addd-6874-4499-9ea1-27ec674bf28a" Nov 24 07:05:48 crc kubenswrapper[4799]: I1124 07:05:48.899748 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-hj8ls" podStartSLOduration=1.957034049 podStartE2EDuration="23.899719834s" podCreationTimestamp="2025-11-24 07:05:25 +0000 UTC" firstStartedPulling="2025-11-24 07:05:26.508265806 +0000 UTC m=+1072.164248280" lastFinishedPulling="2025-11-24 07:05:48.450951591 +0000 UTC m=+1094.106934065" observedRunningTime="2025-11-24 07:05:48.889584968 +0000 UTC m=+1094.545567442" watchObservedRunningTime="2025-11-24 07:05:48.899719834 +0000 UTC m=+1094.555702308" Nov 24 07:05:48 crc kubenswrapper[4799]: I1124 07:05:48.931106 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bdb874957-w6lwg"] Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.019957 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-qf86p"] Nov 24 07:05:49 crc kubenswrapper[4799]: E1124 07:05:49.021027 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03546c3d-2da5-4476-8c22-5a14939cfd80" containerName="glance-db-sync" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.021040 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="03546c3d-2da5-4476-8c22-5a14939cfd80" containerName="glance-db-sync" Nov 24 07:05:49 crc kubenswrapper[4799]: E1124 07:05:49.021087 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048ad289-b5a0-44cf-936f-11680a5dadd9" containerName="init" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.021093 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="048ad289-b5a0-44cf-936f-11680a5dadd9" containerName="init" Nov 24 07:05:49 crc kubenswrapper[4799]: E1124 07:05:49.021118 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048ad289-b5a0-44cf-936f-11680a5dadd9" containerName="dnsmasq-dns" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.021125 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="048ad289-b5a0-44cf-936f-11680a5dadd9" containerName="dnsmasq-dns" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.023155 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="03546c3d-2da5-4476-8c22-5a14939cfd80" containerName="glance-db-sync" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.023210 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="048ad289-b5a0-44cf-936f-11680a5dadd9" containerName="dnsmasq-dns" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.029263 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.048816 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-qf86p"] Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.070413 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-gpvs7" podStartSLOduration=3.002596508 podStartE2EDuration="25.070374942s" podCreationTimestamp="2025-11-24 07:05:24 +0000 UTC" firstStartedPulling="2025-11-24 07:05:26.38151505 +0000 UTC m=+1072.037497524" lastFinishedPulling="2025-11-24 07:05:48.449293474 +0000 UTC m=+1094.105275958" observedRunningTime="2025-11-24 07:05:48.939987503 +0000 UTC m=+1094.595969977" watchObservedRunningTime="2025-11-24 07:05:49.070374942 +0000 UTC m=+1094.726357406" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.087469 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-h5vcf"] Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.096528 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.189759 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-config\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.189829 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlk9w\" (UniqueName: \"kubernetes.io/projected/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-kube-api-access-dlk9w\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.189988 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-sb\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.190044 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-nb\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.190251 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-swift-storage-0\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.190377 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-svc\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.292572 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-nb\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.292664 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-swift-storage-0\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.292717 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-svc\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.292761 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-config\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.292805 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlk9w\" (UniqueName: \"kubernetes.io/projected/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-kube-api-access-dlk9w\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.292836 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-sb\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.293617 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-nb\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.293697 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-swift-storage-0\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.293755 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-sb\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.294569 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-config\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.297436 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-svc\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.323516 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlk9w\" (UniqueName: \"kubernetes.io/projected/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-kube-api-access-dlk9w\") pod \"dnsmasq-dns-76c58b6d97-qf86p\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.390182 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.645973 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="048ad289-b5a0-44cf-936f-11680a5dadd9" path="/var/lib/kubelet/pods/048ad289-b5a0-44cf-936f-11680a5dadd9/volumes" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.824929 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.827034 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.830399 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kg5dw" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.830804 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.831120 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.840935 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.919162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.919207 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-config-data\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.919229 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.919255 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-logs\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.919276 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpl7s\" (UniqueName: \"kubernetes.io/projected/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-kube-api-access-zpl7s\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.919304 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-scripts\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.919347 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.924725 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-qf86p"] Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.929353 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h5vcf" event={"ID":"fe55fcc2-0636-4524-a2f5-8127ab456570","Type":"ContainerStarted","Data":"a14fbfd0d90f8347c1f26eea42fb1a046ab2583fb22623e49ee3a3ffd3956b0c"} Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.929386 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h5vcf" event={"ID":"fe55fcc2-0636-4524-a2f5-8127ab456570","Type":"ContainerStarted","Data":"22eb667d99d87ea812857093cb6150fb4617d841a008b8ffa7e7c98dcdca7052"} Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.931399 4799 generic.go:334] "Generic (PLEG): container finished" podID="52ef5e54-25e2-46c6-bebd-c95be0e4ea0d" containerID="202bb7a523f32b6175c031b0cca7dec206e79cd5edc6497901917d7b8910c247" exitCode=0 Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.932439 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" event={"ID":"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d","Type":"ContainerDied","Data":"202bb7a523f32b6175c031b0cca7dec206e79cd5edc6497901917d7b8910c247"} Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.932465 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" event={"ID":"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d","Type":"ContainerStarted","Data":"ddfa21e5c9a227cc636c27b73ecd45ab7b22ccc5f0fb32023bb6c97f13c9d688"} Nov 24 07:05:49 crc kubenswrapper[4799]: I1124 07:05:49.984807 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-h5vcf" podStartSLOduration=14.984789967 podStartE2EDuration="14.984789967s" podCreationTimestamp="2025-11-24 07:05:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:49.956430305 +0000 UTC m=+1095.612412779" watchObservedRunningTime="2025-11-24 07:05:49.984789967 +0000 UTC m=+1095.640772441" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.020633 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.020765 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.020803 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-config-data\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.020823 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.020896 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-logs\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.020922 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpl7s\" (UniqueName: \"kubernetes.io/projected/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-kube-api-access-zpl7s\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.020977 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-scripts\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.026309 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.026408 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-logs\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.029036 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.030512 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.041570 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-config-data\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.048221 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-scripts\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.068929 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpl7s\" (UniqueName: \"kubernetes.io/projected/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-kube-api-access-zpl7s\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.089218 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.091110 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.091359 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.105796 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.113652 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.157890 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.230988 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.231069 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.231105 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.231175 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-logs\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.231203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.231230 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.231308 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7cqc\" (UniqueName: \"kubernetes.io/projected/b3bc8e56-8446-40ac-8b59-64b9d902aff7-kube-api-access-q7cqc\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.333025 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.333069 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.333130 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-logs\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.333149 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.333169 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.333230 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7cqc\" (UniqueName: \"kubernetes.io/projected/b3bc8e56-8446-40ac-8b59-64b9d902aff7-kube-api-access-q7cqc\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.333268 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.333424 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.334672 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-logs\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.334818 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.338495 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.342976 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-794df4974f-r8fcp" podUID="048ad289-b5a0-44cf-936f-11680a5dadd9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: i/o timeout" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.349132 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.350242 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.355249 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7cqc\" (UniqueName: \"kubernetes.io/projected/b3bc8e56-8446-40ac-8b59-64b9d902aff7-kube-api-access-q7cqc\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.367434 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.401165 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.401601 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:05:50 crc kubenswrapper[4799]: I1124 07:05:50.480058 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:05:51 crc kubenswrapper[4799]: W1124 07:05:51.517822 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5ad083a_71ee_4a11_a298_f0d11d5c2ee2.slice/crio-802871e75b11b117df9577df9680be7cfd37a1f260a97a32ae3aa310b312106a WatchSource:0}: Error finding container 802871e75b11b117df9577df9680be7cfd37a1f260a97a32ae3aa310b312106a: Status 404 returned error can't find the container with id 802871e75b11b117df9577df9680be7cfd37a1f260a97a32ae3aa310b312106a Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.723183 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.789582 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-svc\") pod \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.789908 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-sb\") pod \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.789964 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-nb\") pod \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.790093 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-swift-storage-0\") pod \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.790142 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpk4q\" (UniqueName: \"kubernetes.io/projected/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-kube-api-access-jpk4q\") pod \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.790162 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-config\") pod \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\" (UID: \"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d\") " Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.804839 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-kube-api-access-jpk4q" (OuterVolumeSpecName: "kube-api-access-jpk4q") pod "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d" (UID: "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d"). InnerVolumeSpecName "kube-api-access-jpk4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.867781 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d" (UID: "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.870031 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d" (UID: "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.880191 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d" (UID: "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.887025 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d" (UID: "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.892952 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.892987 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpk4q\" (UniqueName: \"kubernetes.io/projected/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-kube-api-access-jpk4q\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.893002 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.893016 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.893029 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.894686 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-config" (OuterVolumeSpecName: "config") pod "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d" (UID: "52ef5e54-25e2-46c6-bebd-c95be0e4ea0d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.965005 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71fffc24-2640-46ca-92aa-f5a2926f5864","Type":"ContainerStarted","Data":"ed7ae04b94d539a51ace888edb6b2b66f24a3233cebe9b6ec4b0aa1fc5ed9e04"} Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.974840 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" event={"ID":"52ef5e54-25e2-46c6-bebd-c95be0e4ea0d","Type":"ContainerDied","Data":"ddfa21e5c9a227cc636c27b73ecd45ab7b22ccc5f0fb32023bb6c97f13c9d688"} Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.975087 4799 scope.go:117] "RemoveContainer" containerID="202bb7a523f32b6175c031b0cca7dec206e79cd5edc6497901917d7b8910c247" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.975281 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bdb874957-w6lwg" Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.979205 4799 generic.go:334] "Generic (PLEG): container finished" podID="f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" containerID="b3eebfe32951f10b420bda89c65c47b059d9622ed7ae66627b9c9fa050351d5f" exitCode=0 Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.979251 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" event={"ID":"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2","Type":"ContainerDied","Data":"b3eebfe32951f10b420bda89c65c47b059d9622ed7ae66627b9c9fa050351d5f"} Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.979277 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" event={"ID":"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2","Type":"ContainerStarted","Data":"802871e75b11b117df9577df9680be7cfd37a1f260a97a32ae3aa310b312106a"} Nov 24 07:05:51 crc kubenswrapper[4799]: I1124 07:05:51.995310 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:52 crc kubenswrapper[4799]: I1124 07:05:52.094487 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bdb874957-w6lwg"] Nov 24 07:05:52 crc kubenswrapper[4799]: I1124 07:05:52.121575 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bdb874957-w6lwg"] Nov 24 07:05:52 crc kubenswrapper[4799]: I1124 07:05:52.230568 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:05:52 crc kubenswrapper[4799]: I1124 07:05:52.346160 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:05:52 crc kubenswrapper[4799]: W1124 07:05:52.362590 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3bc8e56_8446_40ac_8b59_64b9d902aff7.slice/crio-28c52cab8ec437191d6e7967579953cc12e1fda832f96937df2d0b1fba643bcd WatchSource:0}: Error finding container 28c52cab8ec437191d6e7967579953cc12e1fda832f96937df2d0b1fba643bcd: Status 404 returned error can't find the container with id 28c52cab8ec437191d6e7967579953cc12e1fda832f96937df2d0b1fba643bcd Nov 24 07:05:52 crc kubenswrapper[4799]: I1124 07:05:52.692167 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:05:52 crc kubenswrapper[4799]: I1124 07:05:52.789046 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:05:53 crc kubenswrapper[4799]: I1124 07:05:53.022754 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1","Type":"ContainerStarted","Data":"73ad6278f604146054bb697f87d156cc29c344d34d34e7f14977a2460d96f04e"} Nov 24 07:05:53 crc kubenswrapper[4799]: I1124 07:05:53.026603 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" event={"ID":"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2","Type":"ContainerStarted","Data":"084f09aacec9a679dbbfa8401ef4adb924185db2745ab177a2e1be9ade6871b6"} Nov 24 07:05:53 crc kubenswrapper[4799]: I1124 07:05:53.027614 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:53 crc kubenswrapper[4799]: I1124 07:05:53.030567 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3bc8e56-8446-40ac-8b59-64b9d902aff7","Type":"ContainerStarted","Data":"28c52cab8ec437191d6e7967579953cc12e1fda832f96937df2d0b1fba643bcd"} Nov 24 07:05:53 crc kubenswrapper[4799]: I1124 07:05:53.066983 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" podStartSLOduration=5.066962738 podStartE2EDuration="5.066962738s" podCreationTimestamp="2025-11-24 07:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:53.062198224 +0000 UTC m=+1098.718180698" watchObservedRunningTime="2025-11-24 07:05:53.066962738 +0000 UTC m=+1098.722945212" Nov 24 07:05:53 crc kubenswrapper[4799]: I1124 07:05:53.663564 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52ef5e54-25e2-46c6-bebd-c95be0e4ea0d" path="/var/lib/kubelet/pods/52ef5e54-25e2-46c6-bebd-c95be0e4ea0d/volumes" Nov 24 07:05:54 crc kubenswrapper[4799]: I1124 07:05:54.042653 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3bc8e56-8446-40ac-8b59-64b9d902aff7","Type":"ContainerStarted","Data":"22a17b553077dc6352b0181414414a8c2199ac10cbf1ed0af1106bdf8d651c4c"} Nov 24 07:05:54 crc kubenswrapper[4799]: I1124 07:05:54.045904 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1","Type":"ContainerStarted","Data":"6eb33aaca37c3349f29d18f3e0a1a8408bea083900c105b841a39b4018081309"} Nov 24 07:05:54 crc kubenswrapper[4799]: I1124 07:05:54.047874 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h5vcf" event={"ID":"fe55fcc2-0636-4524-a2f5-8127ab456570","Type":"ContainerDied","Data":"a14fbfd0d90f8347c1f26eea42fb1a046ab2583fb22623e49ee3a3ffd3956b0c"} Nov 24 07:05:54 crc kubenswrapper[4799]: I1124 07:05:54.048408 4799 generic.go:334] "Generic (PLEG): container finished" podID="fe55fcc2-0636-4524-a2f5-8127ab456570" containerID="a14fbfd0d90f8347c1f26eea42fb1a046ab2583fb22623e49ee3a3ffd3956b0c" exitCode=0 Nov 24 07:05:54 crc kubenswrapper[4799]: I1124 07:05:54.049911 4799 generic.go:334] "Generic (PLEG): container finished" podID="b3cd7765-7004-4a36-8331-029e35639568" containerID="2f8c8bdc7ce403fb12cce9fdf8de873b21760d472a20b313828bee7d09f25e41" exitCode=0 Nov 24 07:05:54 crc kubenswrapper[4799]: I1124 07:05:54.049953 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kjxxs" event={"ID":"b3cd7765-7004-4a36-8331-029e35639568","Type":"ContainerDied","Data":"2f8c8bdc7ce403fb12cce9fdf8de873b21760d472a20b313828bee7d09f25e41"} Nov 24 07:05:54 crc kubenswrapper[4799]: I1124 07:05:54.061454 4799 generic.go:334] "Generic (PLEG): container finished" podID="0ccb00cb-56e5-4588-8c60-88b52e92721e" containerID="ae5034ddd416e340224878800975aaca1a7d4c87688395fae504d9fb486b6005" exitCode=0 Nov 24 07:05:54 crc kubenswrapper[4799]: I1124 07:05:54.061834 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hj8ls" event={"ID":"0ccb00cb-56e5-4588-8c60-88b52e92721e","Type":"ContainerDied","Data":"ae5034ddd416e340224878800975aaca1a7d4c87688395fae504d9fb486b6005"} Nov 24 07:05:55 crc kubenswrapper[4799]: I1124 07:05:55.072627 4799 generic.go:334] "Generic (PLEG): container finished" podID="57aab8d2-548c-47cf-9beb-bbe3958ebce6" containerID="34b20cfd09f348a8c1c28edcf64d822d884d0adade4ab63b86fe2ea2d6edf26d" exitCode=0 Nov 24 07:05:55 crc kubenswrapper[4799]: I1124 07:05:55.072931 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gpvs7" event={"ID":"57aab8d2-548c-47cf-9beb-bbe3958ebce6","Type":"ContainerDied","Data":"34b20cfd09f348a8c1c28edcf64d822d884d0adade4ab63b86fe2ea2d6edf26d"} Nov 24 07:05:55 crc kubenswrapper[4799]: I1124 07:05:55.078256 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3bc8e56-8446-40ac-8b59-64b9d902aff7","Type":"ContainerStarted","Data":"416a5ad4b36b34d8c4904cf7659a3abed27d73bf8c90de5da8b689b9a0a396d4"} Nov 24 07:05:55 crc kubenswrapper[4799]: I1124 07:05:55.078331 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b3bc8e56-8446-40ac-8b59-64b9d902aff7" containerName="glance-log" containerID="cri-o://22a17b553077dc6352b0181414414a8c2199ac10cbf1ed0af1106bdf8d651c4c" gracePeriod=30 Nov 24 07:05:55 crc kubenswrapper[4799]: I1124 07:05:55.078349 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b3bc8e56-8446-40ac-8b59-64b9d902aff7" containerName="glance-httpd" containerID="cri-o://416a5ad4b36b34d8c4904cf7659a3abed27d73bf8c90de5da8b689b9a0a396d4" gracePeriod=30 Nov 24 07:05:55 crc kubenswrapper[4799]: I1124 07:05:55.085884 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" containerName="glance-log" containerID="cri-o://6eb33aaca37c3349f29d18f3e0a1a8408bea083900c105b841a39b4018081309" gracePeriod=30 Nov 24 07:05:55 crc kubenswrapper[4799]: I1124 07:05:55.085967 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1","Type":"ContainerStarted","Data":"5007063acf79ec6dbfdd104e41ab1d988cb4680834db607b3bd42be11717fb26"} Nov 24 07:05:55 crc kubenswrapper[4799]: I1124 07:05:55.086657 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" containerName="glance-httpd" containerID="cri-o://5007063acf79ec6dbfdd104e41ab1d988cb4680834db607b3bd42be11717fb26" gracePeriod=30 Nov 24 07:05:55 crc kubenswrapper[4799]: I1124 07:05:55.121349 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.121329338 podStartE2EDuration="7.121329338s" podCreationTimestamp="2025-11-24 07:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:55.118864298 +0000 UTC m=+1100.774846782" watchObservedRunningTime="2025-11-24 07:05:55.121329338 +0000 UTC m=+1100.777311812" Nov 24 07:05:55 crc kubenswrapper[4799]: I1124 07:05:55.147026 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.147008255 podStartE2EDuration="6.147008255s" podCreationTimestamp="2025-11-24 07:05:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:05:55.143188376 +0000 UTC m=+1100.799170850" watchObservedRunningTime="2025-11-24 07:05:55.147008255 +0000 UTC m=+1100.802990729" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.112776 4799 generic.go:334] "Generic (PLEG): container finished" podID="b3bc8e56-8446-40ac-8b59-64b9d902aff7" containerID="416a5ad4b36b34d8c4904cf7659a3abed27d73bf8c90de5da8b689b9a0a396d4" exitCode=0 Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.112957 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3bc8e56-8446-40ac-8b59-64b9d902aff7","Type":"ContainerDied","Data":"416a5ad4b36b34d8c4904cf7659a3abed27d73bf8c90de5da8b689b9a0a396d4"} Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.115142 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3bc8e56-8446-40ac-8b59-64b9d902aff7","Type":"ContainerDied","Data":"22a17b553077dc6352b0181414414a8c2199ac10cbf1ed0af1106bdf8d651c4c"} Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.113404 4799 generic.go:334] "Generic (PLEG): container finished" podID="b3bc8e56-8446-40ac-8b59-64b9d902aff7" containerID="22a17b553077dc6352b0181414414a8c2199ac10cbf1ed0af1106bdf8d651c4c" exitCode=143 Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.117260 4799 generic.go:334] "Generic (PLEG): container finished" podID="0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" containerID="5007063acf79ec6dbfdd104e41ab1d988cb4680834db607b3bd42be11717fb26" exitCode=0 Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.117299 4799 generic.go:334] "Generic (PLEG): container finished" podID="0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" containerID="6eb33aaca37c3349f29d18f3e0a1a8408bea083900c105b841a39b4018081309" exitCode=143 Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.117368 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1","Type":"ContainerDied","Data":"5007063acf79ec6dbfdd104e41ab1d988cb4680834db607b3bd42be11717fb26"} Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.117428 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1","Type":"ContainerDied","Data":"6eb33aaca37c3349f29d18f3e0a1a8408bea083900c105b841a39b4018081309"} Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.255583 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.263498 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.275708 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412455 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-config-data\") pod \"0ccb00cb-56e5-4588-8c60-88b52e92721e\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412524 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-fernet-keys\") pod \"fe55fcc2-0636-4524-a2f5-8127ab456570\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412572 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjls4\" (UniqueName: \"kubernetes.io/projected/fe55fcc2-0636-4524-a2f5-8127ab456570-kube-api-access-zjls4\") pod \"fe55fcc2-0636-4524-a2f5-8127ab456570\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412595 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z52f7\" (UniqueName: \"kubernetes.io/projected/0ccb00cb-56e5-4588-8c60-88b52e92721e-kube-api-access-z52f7\") pod \"0ccb00cb-56e5-4588-8c60-88b52e92721e\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412638 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-config\") pod \"b3cd7765-7004-4a36-8331-029e35639568\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412661 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-combined-ca-bundle\") pod \"0ccb00cb-56e5-4588-8c60-88b52e92721e\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412716 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-config-data\") pod \"fe55fcc2-0636-4524-a2f5-8127ab456570\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412737 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-combined-ca-bundle\") pod \"b3cd7765-7004-4a36-8331-029e35639568\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412759 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkdr2\" (UniqueName: \"kubernetes.io/projected/b3cd7765-7004-4a36-8331-029e35639568-kube-api-access-tkdr2\") pod \"b3cd7765-7004-4a36-8331-029e35639568\" (UID: \"b3cd7765-7004-4a36-8331-029e35639568\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412834 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ccb00cb-56e5-4588-8c60-88b52e92721e-logs\") pod \"0ccb00cb-56e5-4588-8c60-88b52e92721e\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412910 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-combined-ca-bundle\") pod \"fe55fcc2-0636-4524-a2f5-8127ab456570\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412930 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-scripts\") pod \"fe55fcc2-0636-4524-a2f5-8127ab456570\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.412985 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-scripts\") pod \"0ccb00cb-56e5-4588-8c60-88b52e92721e\" (UID: \"0ccb00cb-56e5-4588-8c60-88b52e92721e\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.413006 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-credential-keys\") pod \"fe55fcc2-0636-4524-a2f5-8127ab456570\" (UID: \"fe55fcc2-0636-4524-a2f5-8127ab456570\") " Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.419859 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ccb00cb-56e5-4588-8c60-88b52e92721e-logs" (OuterVolumeSpecName: "logs") pod "0ccb00cb-56e5-4588-8c60-88b52e92721e" (UID: "0ccb00cb-56e5-4588-8c60-88b52e92721e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.475476 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-scripts" (OuterVolumeSpecName: "scripts") pod "fe55fcc2-0636-4524-a2f5-8127ab456570" (UID: "fe55fcc2-0636-4524-a2f5-8127ab456570"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.484039 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fe55fcc2-0636-4524-a2f5-8127ab456570" (UID: "fe55fcc2-0636-4524-a2f5-8127ab456570"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.487147 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3cd7765-7004-4a36-8331-029e35639568-kube-api-access-tkdr2" (OuterVolumeSpecName: "kube-api-access-tkdr2") pod "b3cd7765-7004-4a36-8331-029e35639568" (UID: "b3cd7765-7004-4a36-8331-029e35639568"). InnerVolumeSpecName "kube-api-access-tkdr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.495604 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ccb00cb-56e5-4588-8c60-88b52e92721e-kube-api-access-z52f7" (OuterVolumeSpecName: "kube-api-access-z52f7") pod "0ccb00cb-56e5-4588-8c60-88b52e92721e" (UID: "0ccb00cb-56e5-4588-8c60-88b52e92721e"). InnerVolumeSpecName "kube-api-access-z52f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.495722 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe55fcc2-0636-4524-a2f5-8127ab456570-kube-api-access-zjls4" (OuterVolumeSpecName: "kube-api-access-zjls4") pod "fe55fcc2-0636-4524-a2f5-8127ab456570" (UID: "fe55fcc2-0636-4524-a2f5-8127ab456570"). InnerVolumeSpecName "kube-api-access-zjls4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.501095 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fe55fcc2-0636-4524-a2f5-8127ab456570" (UID: "fe55fcc2-0636-4524-a2f5-8127ab456570"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.501877 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-scripts" (OuterVolumeSpecName: "scripts") pod "0ccb00cb-56e5-4588-8c60-88b52e92721e" (UID: "0ccb00cb-56e5-4588-8c60-88b52e92721e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.510889 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe55fcc2-0636-4524-a2f5-8127ab456570" (UID: "fe55fcc2-0636-4524-a2f5-8127ab456570"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.517076 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ccb00cb-56e5-4588-8c60-88b52e92721e" (UID: "0ccb00cb-56e5-4588-8c60-88b52e92721e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.518025 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ccb00cb-56e5-4588-8c60-88b52e92721e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.518054 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.518063 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.518072 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.518080 4799 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.518088 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.518096 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjls4\" (UniqueName: \"kubernetes.io/projected/fe55fcc2-0636-4524-a2f5-8127ab456570-kube-api-access-zjls4\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.518105 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z52f7\" (UniqueName: \"kubernetes.io/projected/0ccb00cb-56e5-4588-8c60-88b52e92721e-kube-api-access-z52f7\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.518113 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.518120 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkdr2\" (UniqueName: \"kubernetes.io/projected/b3cd7765-7004-4a36-8331-029e35639568-kube-api-access-tkdr2\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.547432 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-config-data" (OuterVolumeSpecName: "config-data") pod "0ccb00cb-56e5-4588-8c60-88b52e92721e" (UID: "0ccb00cb-56e5-4588-8c60-88b52e92721e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.602518 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3cd7765-7004-4a36-8331-029e35639568" (UID: "b3cd7765-7004-4a36-8331-029e35639568"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.613172 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-config-data" (OuterVolumeSpecName: "config-data") pod "fe55fcc2-0636-4524-a2f5-8127ab456570" (UID: "fe55fcc2-0636-4524-a2f5-8127ab456570"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.620044 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ccb00cb-56e5-4588-8c60-88b52e92721e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.620073 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.620089 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe55fcc2-0636-4524-a2f5-8127ab456570-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.625284 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-config" (OuterVolumeSpecName: "config") pod "b3cd7765-7004-4a36-8331-029e35639568" (UID: "b3cd7765-7004-4a36-8331-029e35639568"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:05:56 crc kubenswrapper[4799]: I1124 07:05:56.721153 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b3cd7765-7004-4a36-8331-029e35639568-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.127421 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kjxxs" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.127424 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kjxxs" event={"ID":"b3cd7765-7004-4a36-8331-029e35639568","Type":"ContainerDied","Data":"73f84e58d77d5f3e78ee96d30231e9e72378cb9a4a6b3421b1ce97a433141013"} Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.127491 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73f84e58d77d5f3e78ee96d30231e9e72378cb9a4a6b3421b1ce97a433141013" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.131784 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hj8ls" event={"ID":"0ccb00cb-56e5-4588-8c60-88b52e92721e","Type":"ContainerDied","Data":"ba24fd7e0bdaa34c711156e9f57931de002441ec120ede0972f000b6f0099667"} Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.131865 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba24fd7e0bdaa34c711156e9f57931de002441ec120ede0972f000b6f0099667" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.131949 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hj8ls" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.136232 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h5vcf" event={"ID":"fe55fcc2-0636-4524-a2f5-8127ab456570","Type":"ContainerDied","Data":"22eb667d99d87ea812857093cb6150fb4617d841a008b8ffa7e7c98dcdca7052"} Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.136276 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22eb667d99d87ea812857093cb6150fb4617d841a008b8ffa7e7c98dcdca7052" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.136351 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h5vcf" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.466550 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7b94787bcf-mts4z"] Nov 24 07:05:57 crc kubenswrapper[4799]: E1124 07:05:57.466915 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3cd7765-7004-4a36-8331-029e35639568" containerName="neutron-db-sync" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.466933 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3cd7765-7004-4a36-8331-029e35639568" containerName="neutron-db-sync" Nov 24 07:05:57 crc kubenswrapper[4799]: E1124 07:05:57.466956 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52ef5e54-25e2-46c6-bebd-c95be0e4ea0d" containerName="init" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.466964 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="52ef5e54-25e2-46c6-bebd-c95be0e4ea0d" containerName="init" Nov 24 07:05:57 crc kubenswrapper[4799]: E1124 07:05:57.466989 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe55fcc2-0636-4524-a2f5-8127ab456570" containerName="keystone-bootstrap" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.466996 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe55fcc2-0636-4524-a2f5-8127ab456570" containerName="keystone-bootstrap" Nov 24 07:05:57 crc kubenswrapper[4799]: E1124 07:05:57.467006 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ccb00cb-56e5-4588-8c60-88b52e92721e" containerName="placement-db-sync" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.467013 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ccb00cb-56e5-4588-8c60-88b52e92721e" containerName="placement-db-sync" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.467176 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3cd7765-7004-4a36-8331-029e35639568" containerName="neutron-db-sync" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.467188 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe55fcc2-0636-4524-a2f5-8127ab456570" containerName="keystone-bootstrap" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.467201 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ccb00cb-56e5-4588-8c60-88b52e92721e" containerName="placement-db-sync" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.467210 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="52ef5e54-25e2-46c6-bebd-c95be0e4ea0d" containerName="init" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.467885 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.478909 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-65f6685878-7bjdh"] Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.479329 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kggcp" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.479387 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.479554 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.479674 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.479794 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.480268 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.489301 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.493201 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.493245 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.493423 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.493512 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.493209 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xtntc" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.497243 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7b94787bcf-mts4z"] Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.512122 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65f6685878-7bjdh"] Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544340 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-combined-ca-bundle\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544381 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-scripts\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544420 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-scripts\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544454 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/897d0d90-bf27-4d5b-b136-863379656f34-logs\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544471 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-config-data\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544498 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr8lx\" (UniqueName: \"kubernetes.io/projected/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-kube-api-access-lr8lx\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544519 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-config-data\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544552 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-combined-ca-bundle\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544569 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-public-tls-certs\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544595 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-public-tls-certs\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544612 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-internal-tls-certs\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544635 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-internal-tls-certs\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544659 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s88fq\" (UniqueName: \"kubernetes.io/projected/897d0d90-bf27-4d5b-b136-863379656f34-kube-api-access-s88fq\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544680 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-credential-keys\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.544701 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-fernet-keys\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.577383 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-qf86p"] Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.577601 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" podUID="f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" containerName="dnsmasq-dns" containerID="cri-o://084f09aacec9a679dbbfa8401ef4adb924185db2745ab177a2e1be9ade6871b6" gracePeriod=10 Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.591166 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.605208 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-qnvzb"] Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.606621 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646311 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-config-data\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646358 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-nb\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-combined-ca-bundle\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646438 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-public-tls-certs\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646475 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-public-tls-certs\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646495 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-internal-tls-certs\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646515 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-internal-tls-certs\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646539 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-config\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646557 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s88fq\" (UniqueName: \"kubernetes.io/projected/897d0d90-bf27-4d5b-b136-863379656f34-kube-api-access-s88fq\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646578 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-swift-storage-0\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646596 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-credential-keys\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646618 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-fernet-keys\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646641 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-combined-ca-bundle\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646658 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-scripts\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646680 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-svc\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646698 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhvnl\" (UniqueName: \"kubernetes.io/projected/1e08a70d-d76c-423d-a339-f57a1c879df0-kube-api-access-mhvnl\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646719 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-sb\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646738 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-scripts\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646766 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/897d0d90-bf27-4d5b-b136-863379656f34-logs\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646784 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-config-data\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.646811 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr8lx\" (UniqueName: \"kubernetes.io/projected/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-kube-api-access-lr8lx\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.647515 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-qnvzb"] Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.655804 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-public-tls-certs\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.657450 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-internal-tls-certs\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.657903 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-combined-ca-bundle\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.658187 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/897d0d90-bf27-4d5b-b136-863379656f34-logs\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.663182 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-scripts\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.663303 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-scripts\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.679189 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-fernet-keys\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.682379 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-credential-keys\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.683086 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-config-data\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.683565 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-config-data\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.684753 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-internal-tls-certs\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.685363 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-combined-ca-bundle\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.687069 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s88fq\" (UniqueName: \"kubernetes.io/projected/897d0d90-bf27-4d5b-b136-863379656f34-kube-api-access-s88fq\") pod \"placement-65f6685878-7bjdh\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.694420 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-public-tls-certs\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.695080 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr8lx\" (UniqueName: \"kubernetes.io/projected/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-kube-api-access-lr8lx\") pod \"keystone-7b94787bcf-mts4z\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.748533 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-svc\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.748579 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhvnl\" (UniqueName: \"kubernetes.io/projected/1e08a70d-d76c-423d-a339-f57a1c879df0-kube-api-access-mhvnl\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.748607 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-sb\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.748672 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-nb\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.748737 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-config\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.748758 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-swift-storage-0\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.749597 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-swift-storage-0\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.750107 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-svc\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.750832 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-sb\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.751321 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-nb\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.751810 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-config\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.774960 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhvnl\" (UniqueName: \"kubernetes.io/projected/1e08a70d-d76c-423d-a339-f57a1c879df0-kube-api-access-mhvnl\") pod \"dnsmasq-dns-6c654c9745-qnvzb\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.794837 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.819593 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.854523 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-78cf669d8b-gv6xt"] Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.856885 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.860654 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rsrhk" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.860887 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.862179 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.862364 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.871815 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78cf669d8b-gv6xt"] Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.936506 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.953035 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-config\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.953088 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-combined-ca-bundle\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.953139 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-httpd-config\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.953258 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6kqr\" (UniqueName: \"kubernetes.io/projected/138758e5-99ec-4730-85ec-393117cae96f-kube-api-access-f6kqr\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:57 crc kubenswrapper[4799]: I1124 07:05:57.953304 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-ovndb-tls-certs\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.054664 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6kqr\" (UniqueName: \"kubernetes.io/projected/138758e5-99ec-4730-85ec-393117cae96f-kube-api-access-f6kqr\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.054727 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-ovndb-tls-certs\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.054783 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-config\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.054801 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-combined-ca-bundle\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.054837 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-httpd-config\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.059511 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-combined-ca-bundle\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.059680 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-config\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.059796 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-httpd-config\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.063323 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-ovndb-tls-certs\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.076179 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6kqr\" (UniqueName: \"kubernetes.io/projected/138758e5-99ec-4730-85ec-393117cae96f-kube-api-access-f6kqr\") pod \"neutron-78cf669d8b-gv6xt\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.146030 4799 generic.go:334] "Generic (PLEG): container finished" podID="f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" containerID="084f09aacec9a679dbbfa8401ef4adb924185db2745ab177a2e1be9ade6871b6" exitCode=0 Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.146091 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" event={"ID":"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2","Type":"ContainerDied","Data":"084f09aacec9a679dbbfa8401ef4adb924185db2745ab177a2e1be9ade6871b6"} Nov 24 07:05:58 crc kubenswrapper[4799]: I1124 07:05:58.191913 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:05:59 crc kubenswrapper[4799]: I1124 07:05:59.391132 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" podUID="f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: connect: connection refused" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.043888 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-77fc689cdf-xj7m6"] Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.045409 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.051195 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.051479 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.064251 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77fc689cdf-xj7m6"] Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.091899 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-ovndb-tls-certs\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.091969 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-config\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.091999 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-internal-tls-certs\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.092045 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-httpd-config\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.092235 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-public-tls-certs\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.092333 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-combined-ca-bundle\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.092403 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjff2\" (UniqueName: \"kubernetes.io/projected/0a50ddac-d55d-4e99-8dea-8519b8d732d6-kube-api-access-qjff2\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.194671 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-combined-ca-bundle\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.194740 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjff2\" (UniqueName: \"kubernetes.io/projected/0a50ddac-d55d-4e99-8dea-8519b8d732d6-kube-api-access-qjff2\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.194829 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-ovndb-tls-certs\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.194873 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-config\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.194903 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-internal-tls-certs\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.194944 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-httpd-config\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.194968 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-public-tls-certs\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.201725 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-combined-ca-bundle\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.203783 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-internal-tls-certs\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.204059 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-httpd-config\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.204750 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-public-tls-certs\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.205518 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-config\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.207914 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-ovndb-tls-certs\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.222657 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjff2\" (UniqueName: \"kubernetes.io/projected/0a50ddac-d55d-4e99-8dea-8519b8d732d6-kube-api-access-qjff2\") pod \"neutron-77fc689cdf-xj7m6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.369519 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.678071 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.692370 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.708988 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-logs\") pod \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.709254 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-db-sync-config-data\") pod \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.709358 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-httpd-run\") pod \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.709509 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpl7s\" (UniqueName: \"kubernetes.io/projected/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-kube-api-access-zpl7s\") pod \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.709595 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-combined-ca-bundle\") pod \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.709790 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqx9j\" (UniqueName: \"kubernetes.io/projected/57aab8d2-548c-47cf-9beb-bbe3958ebce6-kube-api-access-vqx9j\") pod \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.709954 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-config-data\") pod \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.710099 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.710201 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-scripts\") pod \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\" (UID: \"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.710300 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-combined-ca-bundle\") pod \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\" (UID: \"57aab8d2-548c-47cf-9beb-bbe3958ebce6\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.710392 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" (UID: "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.710572 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-logs" (OuterVolumeSpecName: "logs") pod "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" (UID: "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.711295 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.711398 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.721564 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.724234 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-scripts" (OuterVolumeSpecName: "scripts") pod "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" (UID: "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.724333 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" (UID: "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.725387 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57aab8d2-548c-47cf-9beb-bbe3958ebce6-kube-api-access-vqx9j" (OuterVolumeSpecName: "kube-api-access-vqx9j") pod "57aab8d2-548c-47cf-9beb-bbe3958ebce6" (UID: "57aab8d2-548c-47cf-9beb-bbe3958ebce6"). InnerVolumeSpecName "kube-api-access-vqx9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.739170 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-kube-api-access-zpl7s" (OuterVolumeSpecName: "kube-api-access-zpl7s") pod "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" (UID: "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1"). InnerVolumeSpecName "kube-api-access-zpl7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.750450 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "57aab8d2-548c-47cf-9beb-bbe3958ebce6" (UID: "57aab8d2-548c-47cf-9beb-bbe3958ebce6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.781533 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" (UID: "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.792623 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57aab8d2-548c-47cf-9beb-bbe3958ebce6" (UID: "57aab8d2-548c-47cf-9beb-bbe3958ebce6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.815464 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-scripts\") pod \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.815527 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7cqc\" (UniqueName: \"kubernetes.io/projected/b3bc8e56-8446-40ac-8b59-64b9d902aff7-kube-api-access-q7cqc\") pod \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.815640 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.815677 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-combined-ca-bundle\") pod \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.816130 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-httpd-run\") pod \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.816193 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-config-data\") pod \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.816333 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-logs\") pod \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\" (UID: \"b3bc8e56-8446-40ac-8b59-64b9d902aff7\") " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.817160 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqx9j\" (UniqueName: \"kubernetes.io/projected/57aab8d2-548c-47cf-9beb-bbe3958ebce6-kube-api-access-vqx9j\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.817196 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.817206 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.817216 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.817228 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57aab8d2-548c-47cf-9beb-bbe3958ebce6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.817238 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpl7s\" (UniqueName: \"kubernetes.io/projected/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-kube-api-access-zpl7s\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.817247 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.827604 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b3bc8e56-8446-40ac-8b59-64b9d902aff7" (UID: "b3bc8e56-8446-40ac-8b59-64b9d902aff7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.828307 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-logs" (OuterVolumeSpecName: "logs") pod "b3bc8e56-8446-40ac-8b59-64b9d902aff7" (UID: "b3bc8e56-8446-40ac-8b59-64b9d902aff7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.832769 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-scripts" (OuterVolumeSpecName: "scripts") pod "b3bc8e56-8446-40ac-8b59-64b9d902aff7" (UID: "b3bc8e56-8446-40ac-8b59-64b9d902aff7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.834904 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "b3bc8e56-8446-40ac-8b59-64b9d902aff7" (UID: "b3bc8e56-8446-40ac-8b59-64b9d902aff7"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.841652 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3bc8e56-8446-40ac-8b59-64b9d902aff7-kube-api-access-q7cqc" (OuterVolumeSpecName: "kube-api-access-q7cqc") pod "b3bc8e56-8446-40ac-8b59-64b9d902aff7" (UID: "b3bc8e56-8446-40ac-8b59-64b9d902aff7"). InnerVolumeSpecName "kube-api-access-q7cqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.884823 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-config-data" (OuterVolumeSpecName: "config-data") pod "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" (UID: "0d0ca8c3-bed5-49e9-96b1-1834fa104bd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.886200 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.904642 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3bc8e56-8446-40ac-8b59-64b9d902aff7" (UID: "b3bc8e56-8446-40ac-8b59-64b9d902aff7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.918924 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.918961 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7cqc\" (UniqueName: \"kubernetes.io/projected/b3bc8e56-8446-40ac-8b59-64b9d902aff7-kube-api-access-q7cqc\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.918994 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.919004 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.919012 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.919021 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.919029 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3bc8e56-8446-40ac-8b59-64b9d902aff7-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.919038 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.940894 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-config-data" (OuterVolumeSpecName: "config-data") pod "b3bc8e56-8446-40ac-8b59-64b9d902aff7" (UID: "b3bc8e56-8446-40ac-8b59-64b9d902aff7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.942251 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 24 07:06:00 crc kubenswrapper[4799]: I1124 07:06:00.956988 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.024531 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-sb\") pod \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.024605 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlk9w\" (UniqueName: \"kubernetes.io/projected/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-kube-api-access-dlk9w\") pod \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.024677 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-config\") pod \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.024737 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-swift-storage-0\") pod \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.024892 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-nb\") pod \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.025001 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-svc\") pod \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\" (UID: \"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2\") " Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.025509 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.025525 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3bc8e56-8446-40ac-8b59-64b9d902aff7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.033550 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-kube-api-access-dlk9w" (OuterVolumeSpecName: "kube-api-access-dlk9w") pod "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" (UID: "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2"). InnerVolumeSpecName "kube-api-access-dlk9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.089765 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" (UID: "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.097184 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" (UID: "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.099564 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" (UID: "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.102136 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-config" (OuterVolumeSpecName: "config") pod "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" (UID: "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.113665 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" (UID: "f5ad083a-71ee-4a11-a298-f0d11d5c2ee2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.141659 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.141729 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlk9w\" (UniqueName: \"kubernetes.io/projected/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-kube-api-access-dlk9w\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.141749 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.141760 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.141775 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.141790 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.171048 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" event={"ID":"f5ad083a-71ee-4a11-a298-f0d11d5c2ee2","Type":"ContainerDied","Data":"802871e75b11b117df9577df9680be7cfd37a1f260a97a32ae3aa310b312106a"} Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.171111 4799 scope.go:117] "RemoveContainer" containerID="084f09aacec9a679dbbfa8401ef4adb924185db2745ab177a2e1be9ade6871b6" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.171223 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-qf86p" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.173249 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gpvs7" event={"ID":"57aab8d2-548c-47cf-9beb-bbe3958ebce6","Type":"ContainerDied","Data":"b4e6d1c789d5ad1997ff8598a7fddd4efe69b58699f9faa41cebffae7648a034"} Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.173277 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4e6d1c789d5ad1997ff8598a7fddd4efe69b58699f9faa41cebffae7648a034" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.174074 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gpvs7" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.187473 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71fffc24-2640-46ca-92aa-f5a2926f5864","Type":"ContainerStarted","Data":"239a6f800e16ecbd1a6e367158fc89abbb1733a4c14a5996855ff9fdc8c1b554"} Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.191134 4799 scope.go:117] "RemoveContainer" containerID="b3eebfe32951f10b420bda89c65c47b059d9622ed7ae66627b9c9fa050351d5f" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.191382 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.198058 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3bc8e56-8446-40ac-8b59-64b9d902aff7","Type":"ContainerDied","Data":"28c52cab8ec437191d6e7967579953cc12e1fda832f96937df2d0b1fba643bcd"} Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.203143 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d0ca8c3-bed5-49e9-96b1-1834fa104bd1","Type":"ContainerDied","Data":"73ad6278f604146054bb697f87d156cc29c344d34d34e7f14977a2460d96f04e"} Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.203269 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.251339 4799 scope.go:117] "RemoveContainer" containerID="416a5ad4b36b34d8c4904cf7659a3abed27d73bf8c90de5da8b689b9a0a396d4" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.385809 4799 scope.go:117] "RemoveContainer" containerID="22a17b553077dc6352b0181414414a8c2199ac10cbf1ed0af1106bdf8d651c4c" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.404783 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.426837 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.430185 4799 scope.go:117] "RemoveContainer" containerID="5007063acf79ec6dbfdd104e41ab1d988cb4680834db607b3bd42be11717fb26" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.439926 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.455967 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474119 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:06:01 crc kubenswrapper[4799]: E1124 07:06:01.474518 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3bc8e56-8446-40ac-8b59-64b9d902aff7" containerName="glance-httpd" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474537 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3bc8e56-8446-40ac-8b59-64b9d902aff7" containerName="glance-httpd" Nov 24 07:06:01 crc kubenswrapper[4799]: E1124 07:06:01.474549 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" containerName="init" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474556 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" containerName="init" Nov 24 07:06:01 crc kubenswrapper[4799]: E1124 07:06:01.474564 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" containerName="dnsmasq-dns" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474570 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" containerName="dnsmasq-dns" Nov 24 07:06:01 crc kubenswrapper[4799]: E1124 07:06:01.474581 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57aab8d2-548c-47cf-9beb-bbe3958ebce6" containerName="barbican-db-sync" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474587 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="57aab8d2-548c-47cf-9beb-bbe3958ebce6" containerName="barbican-db-sync" Nov 24 07:06:01 crc kubenswrapper[4799]: E1124 07:06:01.474598 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" containerName="glance-httpd" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474604 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" containerName="glance-httpd" Nov 24 07:06:01 crc kubenswrapper[4799]: E1124 07:06:01.474619 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" containerName="glance-log" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474626 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" containerName="glance-log" Nov 24 07:06:01 crc kubenswrapper[4799]: E1124 07:06:01.474656 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3bc8e56-8446-40ac-8b59-64b9d902aff7" containerName="glance-log" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474662 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3bc8e56-8446-40ac-8b59-64b9d902aff7" containerName="glance-log" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474826 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3bc8e56-8446-40ac-8b59-64b9d902aff7" containerName="glance-httpd" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474835 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" containerName="glance-httpd" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474866 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3bc8e56-8446-40ac-8b59-64b9d902aff7" containerName="glance-log" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474875 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" containerName="glance-log" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474895 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" containerName="dnsmasq-dns" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.474905 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="57aab8d2-548c-47cf-9beb-bbe3958ebce6" containerName="barbican-db-sync" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.475794 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.481552 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.481735 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.481980 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kg5dw" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.482171 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.490139 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.492644 4799 scope.go:117] "RemoveContainer" containerID="6eb33aaca37c3349f29d18f3e0a1a8408bea083900c105b841a39b4018081309" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.514534 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.516076 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.519188 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.519252 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.533353 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-qf86p"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.543918 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-qf86p"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.551146 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.551181 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-logs\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.551205 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-scripts\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.551479 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.551506 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.551530 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7blll\" (UniqueName: \"kubernetes.io/projected/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-kube-api-access-7blll\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.551555 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-config-data\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.551576 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.554387 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.561020 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-qnvzb"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.576397 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65f6685878-7bjdh"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.583347 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7b94787bcf-mts4z"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.642923 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d0ca8c3-bed5-49e9-96b1-1834fa104bd1" path="/var/lib/kubelet/pods/0d0ca8c3-bed5-49e9-96b1-1834fa104bd1/volumes" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.643604 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3bc8e56-8446-40ac-8b59-64b9d902aff7" path="/var/lib/kubelet/pods/b3bc8e56-8446-40ac-8b59-64b9d902aff7/volumes" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.644321 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5ad083a-71ee-4a11-a298-f0d11d5c2ee2" path="/var/lib/kubelet/pods/f5ad083a-71ee-4a11-a298-f0d11d5c2ee2/volumes" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656305 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7blll\" (UniqueName: \"kubernetes.io/projected/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-kube-api-access-7blll\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656387 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-config-data\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656419 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glvxs\" (UniqueName: \"kubernetes.io/projected/f7d704f4-3e1b-4d87-807c-603b77175af5-kube-api-access-glvxs\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656449 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656469 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656488 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656650 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656673 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656698 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-logs\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656724 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656756 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-scripts\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656799 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.656934 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.657021 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.657066 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.657107 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.658021 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.660332 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-logs\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.664283 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.665213 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.668173 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.674144 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-scripts\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.675051 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-config-data\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.675481 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78cf669d8b-gv6xt"] Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.688560 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7blll\" (UniqueName: \"kubernetes.io/projected/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-kube-api-access-7blll\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.721822 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " pod="openstack/glance-default-external-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.763421 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.763480 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.763532 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.763616 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.763683 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.763745 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glvxs\" (UniqueName: \"kubernetes.io/projected/f7d704f4-3e1b-4d87-807c-603b77175af5-kube-api-access-glvxs\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.763762 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.763780 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.765021 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.774343 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.774595 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.786709 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.804662 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.805337 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.807315 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.865009 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glvxs\" (UniqueName: \"kubernetes.io/projected/f7d704f4-3e1b-4d87-807c-603b77175af5-kube-api-access-glvxs\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.950339 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:06:01 crc kubenswrapper[4799]: I1124 07:06:01.968810 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.002245 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.096863 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7dd777b755-wgf6d"] Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.098348 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.105814 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.109287 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.109466 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d296j" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.139256 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6fbb66bcc8-62525"] Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.149186 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.153151 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.158388 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6fbb66bcc8-62525"] Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.215214 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9ee9f9f-66be-467a-a982-9ede1e1b682c-logs\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.215294 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-combined-ca-bundle\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.215351 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.215540 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68frq\" (UniqueName: \"kubernetes.io/projected/a9ee9f9f-66be-467a-a982-9ede1e1b682c-kube-api-access-68frq\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.215585 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data-custom\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.290503 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7dd777b755-wgf6d"] Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.330386 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.330487 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68frq\" (UniqueName: \"kubernetes.io/projected/a9ee9f9f-66be-467a-a982-9ede1e1b682c-kube-api-access-68frq\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.330531 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89r6q\" (UniqueName: \"kubernetes.io/projected/c986e5b7-9864-46ae-9c99-9072e9c6ce34-kube-api-access-89r6q\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.330561 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data-custom\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.330650 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c986e5b7-9864-46ae-9c99-9072e9c6ce34-logs\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.330685 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9ee9f9f-66be-467a-a982-9ede1e1b682c-logs\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.330724 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-combined-ca-bundle\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.330777 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.330905 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-combined-ca-bundle\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.331042 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data-custom\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.333217 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9ee9f9f-66be-467a-a982-9ede1e1b682c-logs\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.386668 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-combined-ca-bundle\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.392748 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68frq\" (UniqueName: \"kubernetes.io/projected/a9ee9f9f-66be-467a-a982-9ede1e1b682c-kube-api-access-68frq\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.393816 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data-custom\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.434091 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-combined-ca-bundle\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.434434 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data-custom\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.434475 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.434515 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89r6q\" (UniqueName: \"kubernetes.io/projected/c986e5b7-9864-46ae-9c99-9072e9c6ce34-kube-api-access-89r6q\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.434590 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c986e5b7-9864-46ae-9c99-9072e9c6ce34-logs\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.438973 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-qnvzb"] Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.447067 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65f6685878-7bjdh" event={"ID":"897d0d90-bf27-4d5b-b136-863379656f34","Type":"ContainerStarted","Data":"702c77d5a08ac68b078c369149870d0f8a51352018417fd5ffd7488783fd889b"} Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.447123 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65f6685878-7bjdh" event={"ID":"897d0d90-bf27-4d5b-b136-863379656f34","Type":"ContainerStarted","Data":"f1c6fcbd0f68b00e30c3cea438c780a1b0f1f06e461cd0eae3ef252321de1f0e"} Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.450072 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data\") pod \"barbican-worker-7dd777b755-wgf6d\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.459509 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-combined-ca-bundle\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.459580 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-nwn2f"] Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.461434 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.464294 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c986e5b7-9864-46ae-9c99-9072e9c6ce34-logs\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.472259 4799 generic.go:334] "Generic (PLEG): container finished" podID="1e08a70d-d76c-423d-a339-f57a1c879df0" containerID="6eed7168302b8f10040d414474949ba7667211e64eb91c82b43bd8b08401b6de" exitCode=0 Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.472729 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" event={"ID":"1e08a70d-d76c-423d-a339-f57a1c879df0","Type":"ContainerDied","Data":"6eed7168302b8f10040d414474949ba7667211e64eb91c82b43bd8b08401b6de"} Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.472790 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" event={"ID":"1e08a70d-d76c-423d-a339-f57a1c879df0","Type":"ContainerStarted","Data":"51395a7971d38d87beaa875374e94aab019ae4ed1c8022d6dc2213fc4e322b36"} Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.473140 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.480443 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data-custom\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.483805 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-nwn2f"] Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.498258 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5d7959f94d-skwn6"] Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.499971 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.502408 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89r6q\" (UniqueName: \"kubernetes.io/projected/c986e5b7-9864-46ae-9c99-9072e9c6ce34-kube-api-access-89r6q\") pod \"barbican-keystone-listener-6fbb66bcc8-62525\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.507366 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.508027 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.522828 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.527397 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b94787bcf-mts4z" event={"ID":"7603a96c-19e2-48ce-9c5f-69eeacc8cd85","Type":"ContainerStarted","Data":"60197a8ca0e4d3646b4e4bd496c61312a49d9782b200ab25fbbb3989c5ad162d"} Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.527587 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.536439 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh9ks\" (UniqueName: \"kubernetes.io/projected/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-kube-api-access-gh9ks\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.536515 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.536538 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.536564 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-config\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.536611 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-combined-ca-bundle\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.536637 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v2x9\" (UniqueName: \"kubernetes.io/projected/f633fe14-fd8c-4738-9960-4fff9619c027-kube-api-access-5v2x9\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.536665 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data-custom\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.536696 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f633fe14-fd8c-4738-9960-4fff9619c027-logs\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.536755 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.536772 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.536791 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-svc\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.540742 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cf669d8b-gv6xt" event={"ID":"138758e5-99ec-4730-85ec-393117cae96f","Type":"ContainerStarted","Data":"1e25741d3e9fc033beede695361855121a2308d99f8a2c0df4b1badd97b27c2d"} Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.549530 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d7959f94d-skwn6"] Nov 24 07:06:02 crc kubenswrapper[4799]: E1124 07:06:02.555038 4799 mount_linux.go:282] Mount failed: exit status 32 Nov 24 07:06:02 crc kubenswrapper[4799]: Mounting command: mount Nov 24 07:06:02 crc kubenswrapper[4799]: Mounting arguments: --no-canonicalize -o bind /proc/4799/fd/21 /var/lib/kubelet/pods/1e08a70d-d76c-423d-a339-f57a1c879df0/volume-subpaths/dns-svc/dnsmasq-dns/1 Nov 24 07:06:02 crc kubenswrapper[4799]: Output: mount: /var/lib/kubelet/pods/1e08a70d-d76c-423d-a339-f57a1c879df0/volume-subpaths/dns-svc/dnsmasq-dns/1: mount(2) system call failed: No such file or directory. Nov 24 07:06:02 crc kubenswrapper[4799]: E1124 07:06:02.596442 4799 kubelet_pods.go:349] "Failed to prepare subPath for volumeMount of the container" err=< Nov 24 07:06:02 crc kubenswrapper[4799]: error mounting /var/lib/kubelet/pods/1e08a70d-d76c-423d-a339-f57a1c879df0/volumes/kubernetes.io~configmap/dns-svc/..2025_11_24_07_05_57.1349695278/dns-svc: mount failed: exit status 32 Nov 24 07:06:02 crc kubenswrapper[4799]: Mounting command: mount Nov 24 07:06:02 crc kubenswrapper[4799]: Mounting arguments: --no-canonicalize -o bind /proc/4799/fd/21 /var/lib/kubelet/pods/1e08a70d-d76c-423d-a339-f57a1c879df0/volume-subpaths/dns-svc/dnsmasq-dns/1 Nov 24 07:06:02 crc kubenswrapper[4799]: Output: mount: /var/lib/kubelet/pods/1e08a70d-d76c-423d-a339-f57a1c879df0/volume-subpaths/dns-svc/dnsmasq-dns/1: mount(2) system call failed: No such file or directory. Nov 24 07:06:02 crc kubenswrapper[4799]: > containerName="dnsmasq-dns" volumeMountName="dns-svc" Nov 24 07:06:02 crc kubenswrapper[4799]: E1124 07:06:02.596671 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n587h5c6h66fhc7hcch59hffh5f4h665h5cdh57h68dh66bh5f9hcbh96h6fh578h56dh656h658h76h65ch68fh655h5bbh4h6fh5c5h668h697hfdq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mhvnl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6c654c9745-qnvzb_openstack(1e08a70d-d76c-423d-a339-f57a1c879df0): CreateContainerConfigError: failed to prepare subPath for volumeMount \"dns-svc\" of container \"dnsmasq-dns\"" logger="UnhandledError" Nov 24 07:06:02 crc kubenswrapper[4799]: E1124 07:06:02.597952 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerConfigError: \"failed to prepare subPath for volumeMount \\\"dns-svc\\\" of container \\\"dnsmasq-dns\\\"\"" pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" podUID="1e08a70d-d76c-423d-a339-f57a1c879df0" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.638714 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.639223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.639254 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-config\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.639302 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-combined-ca-bundle\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.639322 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v2x9\" (UniqueName: \"kubernetes.io/projected/f633fe14-fd8c-4738-9960-4fff9619c027-kube-api-access-5v2x9\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.639361 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data-custom\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.639394 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f633fe14-fd8c-4738-9960-4fff9619c027-logs\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.639427 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-svc\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.639445 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.639466 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.639582 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh9ks\" (UniqueName: \"kubernetes.io/projected/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-kube-api-access-gh9ks\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.639959 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f633fe14-fd8c-4738-9960-4fff9619c027-logs\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.640758 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.641324 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-config\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.642112 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.642179 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-svc\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.644043 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.653326 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-combined-ca-bundle\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.664599 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data-custom\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.671680 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh9ks\" (UniqueName: \"kubernetes.io/projected/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-kube-api-access-gh9ks\") pod \"dnsmasq-dns-5cc67f459c-nwn2f\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.672239 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.674925 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77fc689cdf-xj7m6"] Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.679273 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v2x9\" (UniqueName: \"kubernetes.io/projected/f633fe14-fd8c-4738-9960-4fff9619c027-kube-api-access-5v2x9\") pod \"barbican-api-5d7959f94d-skwn6\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.690556 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7b94787bcf-mts4z" podStartSLOduration=5.690531168 podStartE2EDuration="5.690531168s" podCreationTimestamp="2025-11-24 07:05:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:02.596483979 +0000 UTC m=+1108.252466453" watchObservedRunningTime="2025-11-24 07:06:02.690531168 +0000 UTC m=+1108.346513642" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.843435 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.900750 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:02 crc kubenswrapper[4799]: I1124 07:06:02.901723 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.094792 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.210032 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7dd777b755-wgf6d"] Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.361934 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6fbb66bcc8-62525"] Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.569403 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7dd777b755-wgf6d" event={"ID":"a9ee9f9f-66be-467a-a982-9ede1e1b682c","Type":"ContainerStarted","Data":"953ffb6c5a3fad664636d7f6e9310636d909840ca49d65fdc75eb8af5c39dd92"} Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.570930 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65f6685878-7bjdh" event={"ID":"897d0d90-bf27-4d5b-b136-863379656f34","Type":"ContainerStarted","Data":"62f862c6ff3823a347d374cd9375930cec2ce5b254235146abb2a0f3a2a1a61c"} Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.572235 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.572262 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.575963 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7d704f4-3e1b-4d87-807c-603b77175af5","Type":"ContainerStarted","Data":"6ee960a759929c1f2b370e54fa7cf51f11c31520e13144513846f8842770e57d"} Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.580246 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77fc689cdf-xj7m6" event={"ID":"0a50ddac-d55d-4e99-8dea-8519b8d732d6","Type":"ContainerStarted","Data":"02ffa0bb3d383e7c692b859a1e3376a9f3f7187ae40a809812908dbb07a5603c"} Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.580307 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77fc689cdf-xj7m6" event={"ID":"0a50ddac-d55d-4e99-8dea-8519b8d732d6","Type":"ContainerStarted","Data":"bab175d3ca6e71c3354bcdb618c71fac89d3d23a05be126cef27869bd259e337"} Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.586841 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" event={"ID":"c986e5b7-9864-46ae-9c99-9072e9c6ce34","Type":"ContainerStarted","Data":"b59051fadd3bd06a2fad5470bc4264fd6214b1856014a42cba1775ec385c56ad"} Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.586958 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-nwn2f"] Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.611431 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-65f6685878-7bjdh" podStartSLOduration=6.611413186 podStartE2EDuration="6.611413186s" podCreationTimestamp="2025-11-24 07:05:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:03.597309677 +0000 UTC m=+1109.253292151" watchObservedRunningTime="2025-11-24 07:06:03.611413186 +0000 UTC m=+1109.267395660" Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.617651 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b94787bcf-mts4z" event={"ID":"7603a96c-19e2-48ce-9c5f-69eeacc8cd85","Type":"ContainerStarted","Data":"29a4295f9247968819fb9ab51fb68ba67eeb2fb13b12fb980c93984bfc24c9e9"} Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.622785 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cf669d8b-gv6xt" event={"ID":"138758e5-99ec-4730-85ec-393117cae96f","Type":"ContainerStarted","Data":"f0d526cab039dc7051e8feec957eed77f1379399fb68a6b9d251d359457ebd40"} Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.622825 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cf669d8b-gv6xt" event={"ID":"138758e5-99ec-4730-85ec-393117cae96f","Type":"ContainerStarted","Data":"3259793339fb0fed04151354af63bbc8f62939707a77cc329a87cb86d9871b65"} Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.626219 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.727547 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b6ec1b6c-dd0f-43e0-8894-d07fee864da5","Type":"ContainerStarted","Data":"acde285dbf829486a21a0e6e3a1134310a9ba803f0623ff5a73264f5b41cd2b9"} Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.868322 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-78cf669d8b-gv6xt" podStartSLOduration=6.868299142 podStartE2EDuration="6.868299142s" podCreationTimestamp="2025-11-24 07:05:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:03.658549199 +0000 UTC m=+1109.314531683" watchObservedRunningTime="2025-11-24 07:06:03.868299142 +0000 UTC m=+1109.524281616" Nov 24 07:06:03 crc kubenswrapper[4799]: I1124 07:06:03.871273 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d7959f94d-skwn6"] Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.128571 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.295768 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-sb\") pod \"1e08a70d-d76c-423d-a339-f57a1c879df0\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.295810 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-swift-storage-0\") pod \"1e08a70d-d76c-423d-a339-f57a1c879df0\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.295998 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-svc\") pod \"1e08a70d-d76c-423d-a339-f57a1c879df0\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.296027 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-config\") pod \"1e08a70d-d76c-423d-a339-f57a1c879df0\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.296166 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-nb\") pod \"1e08a70d-d76c-423d-a339-f57a1c879df0\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.296229 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhvnl\" (UniqueName: \"kubernetes.io/projected/1e08a70d-d76c-423d-a339-f57a1c879df0-kube-api-access-mhvnl\") pod \"1e08a70d-d76c-423d-a339-f57a1c879df0\" (UID: \"1e08a70d-d76c-423d-a339-f57a1c879df0\") " Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.301035 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e08a70d-d76c-423d-a339-f57a1c879df0-kube-api-access-mhvnl" (OuterVolumeSpecName: "kube-api-access-mhvnl") pod "1e08a70d-d76c-423d-a339-f57a1c879df0" (UID: "1e08a70d-d76c-423d-a339-f57a1c879df0"). InnerVolumeSpecName "kube-api-access-mhvnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.327750 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1e08a70d-d76c-423d-a339-f57a1c879df0" (UID: "1e08a70d-d76c-423d-a339-f57a1c879df0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.331338 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1e08a70d-d76c-423d-a339-f57a1c879df0" (UID: "1e08a70d-d76c-423d-a339-f57a1c879df0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.402180 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhvnl\" (UniqueName: \"kubernetes.io/projected/1e08a70d-d76c-423d-a339-f57a1c879df0-kube-api-access-mhvnl\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.402404 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.402461 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.447669 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1e08a70d-d76c-423d-a339-f57a1c879df0" (UID: "1e08a70d-d76c-423d-a339-f57a1c879df0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.509493 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.529720 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1e08a70d-d76c-423d-a339-f57a1c879df0" (UID: "1e08a70d-d76c-423d-a339-f57a1c879df0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.532339 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-config" (OuterVolumeSpecName: "config") pod "1e08a70d-d76c-423d-a339-f57a1c879df0" (UID: "1e08a70d-d76c-423d-a339-f57a1c879df0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.612030 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.612450 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e08a70d-d76c-423d-a339-f57a1c879df0-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.668162 4799 generic.go:334] "Generic (PLEG): container finished" podID="3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" containerID="acac3cacb1852d0871215875e1702280b7581812d0c21cbcb57e32c22a26256b" exitCode=0 Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.669359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" event={"ID":"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5","Type":"ContainerDied","Data":"acac3cacb1852d0871215875e1702280b7581812d0c21cbcb57e32c22a26256b"} Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.669404 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" event={"ID":"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5","Type":"ContainerStarted","Data":"8874e78da064a4f3d93503720862a02cd78f532d28601ce978e5f593d824ae70"} Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.682070 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7959f94d-skwn6" event={"ID":"f633fe14-fd8c-4738-9960-4fff9619c027","Type":"ContainerStarted","Data":"f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691"} Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.682124 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7959f94d-skwn6" event={"ID":"f633fe14-fd8c-4738-9960-4fff9619c027","Type":"ContainerStarted","Data":"b69ba6812367bc13da1033eb9600fd77dbf01dc68bbbc1ee747a04e869827152"} Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.688988 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b6ec1b6c-dd0f-43e0-8894-d07fee864da5","Type":"ContainerStarted","Data":"4cb319180de5b546f429d89843adb706cbbabec96cc2e0ded3e2f44af9f83e16"} Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.736385 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" event={"ID":"1e08a70d-d76c-423d-a339-f57a1c879df0","Type":"ContainerDied","Data":"51395a7971d38d87beaa875374e94aab019ae4ed1c8022d6dc2213fc4e322b36"} Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.736439 4799 scope.go:117] "RemoveContainer" containerID="6eed7168302b8f10040d414474949ba7667211e64eb91c82b43bd8b08401b6de" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.736551 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-qnvzb" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.761207 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7d704f4-3e1b-4d87-807c-603b77175af5","Type":"ContainerStarted","Data":"49b3ecf627707e4282cef4a93505d13d8a3d2349393427ee26e83ab6dc083bf4"} Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.775950 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77fc689cdf-xj7m6" event={"ID":"0a50ddac-d55d-4e99-8dea-8519b8d732d6","Type":"ContainerStarted","Data":"d3a425e1e004a0c21cb397f3983c5fe55824d01453fd599218ea20167f5095a8"} Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.776730 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.836187 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-qnvzb"] Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.863525 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-qnvzb"] Nov 24 07:06:04 crc kubenswrapper[4799]: I1124 07:06:04.891716 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-77fc689cdf-xj7m6" podStartSLOduration=4.89169466 podStartE2EDuration="4.89169466s" podCreationTimestamp="2025-11-24 07:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:04.838511906 +0000 UTC m=+1110.494494380" watchObservedRunningTime="2025-11-24 07:06:04.89169466 +0000 UTC m=+1110.547677134" Nov 24 07:06:05 crc kubenswrapper[4799]: I1124 07:06:05.645954 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e08a70d-d76c-423d-a339-f57a1c879df0" path="/var/lib/kubelet/pods/1e08a70d-d76c-423d-a339-f57a1c879df0/volumes" Nov 24 07:06:05 crc kubenswrapper[4799]: I1124 07:06:05.784661 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" event={"ID":"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5","Type":"ContainerStarted","Data":"3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb"} Nov 24 07:06:05 crc kubenswrapper[4799]: I1124 07:06:05.785925 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:05 crc kubenswrapper[4799]: I1124 07:06:05.790256 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7959f94d-skwn6" event={"ID":"f633fe14-fd8c-4738-9960-4fff9619c027","Type":"ContainerStarted","Data":"4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d"} Nov 24 07:06:05 crc kubenswrapper[4799]: I1124 07:06:05.791312 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:05 crc kubenswrapper[4799]: I1124 07:06:05.791361 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:05 crc kubenswrapper[4799]: I1124 07:06:05.793035 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b6ec1b6c-dd0f-43e0-8894-d07fee864da5","Type":"ContainerStarted","Data":"f2093df971a210f5c37ee68ffd3cb3d71162fa62ac52dc98027d51155893ca23"} Nov 24 07:06:05 crc kubenswrapper[4799]: I1124 07:06:05.797430 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7d704f4-3e1b-4d87-807c-603b77175af5","Type":"ContainerStarted","Data":"ac1311b99af63466f8fdf3dde035f12ac5ad51bace15d1a69e8e29e4043627ac"} Nov 24 07:06:05 crc kubenswrapper[4799]: I1124 07:06:05.816293 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" podStartSLOduration=3.816275883 podStartE2EDuration="3.816275883s" podCreationTimestamp="2025-11-24 07:06:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:05.805253071 +0000 UTC m=+1111.461235545" watchObservedRunningTime="2025-11-24 07:06:05.816275883 +0000 UTC m=+1111.472258357" Nov 24 07:06:05 crc kubenswrapper[4799]: I1124 07:06:05.847632 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.847606459 podStartE2EDuration="4.847606459s" podCreationTimestamp="2025-11-24 07:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:05.824913717 +0000 UTC m=+1111.480896201" watchObservedRunningTime="2025-11-24 07:06:05.847606459 +0000 UTC m=+1111.503588933" Nov 24 07:06:05 crc kubenswrapper[4799]: I1124 07:06:05.883211 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.883192376 podStartE2EDuration="4.883192376s" podCreationTimestamp="2025-11-24 07:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:05.865013351 +0000 UTC m=+1111.520995815" watchObservedRunningTime="2025-11-24 07:06:05.883192376 +0000 UTC m=+1111.539174840" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.366739 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5d7959f94d-skwn6" podStartSLOduration=4.366717982 podStartE2EDuration="4.366717982s" podCreationTimestamp="2025-11-24 07:06:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:05.900025232 +0000 UTC m=+1111.556007716" watchObservedRunningTime="2025-11-24 07:06:06.366717982 +0000 UTC m=+1112.022700466" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.371717 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-644b86779b-b4psv"] Nov 24 07:06:06 crc kubenswrapper[4799]: E1124 07:06:06.372334 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e08a70d-d76c-423d-a339-f57a1c879df0" containerName="init" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.372437 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e08a70d-d76c-423d-a339-f57a1c879df0" containerName="init" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.373032 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e08a70d-d76c-423d-a339-f57a1c879df0" containerName="init" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.374674 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.377293 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.378624 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.389100 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-644b86779b-b4psv"] Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.465327 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebfe77df-bbcd-412f-9703-78e23d93f410-logs\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.465397 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-internal-tls-certs\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.465592 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.465643 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h44k\" (UniqueName: \"kubernetes.io/projected/ebfe77df-bbcd-412f-9703-78e23d93f410-kube-api-access-8h44k\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.465775 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-public-tls-certs\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.466023 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-combined-ca-bundle\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.466264 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data-custom\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.567941 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data-custom\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.568005 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebfe77df-bbcd-412f-9703-78e23d93f410-logs\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.568041 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-internal-tls-certs\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.568083 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.568105 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h44k\" (UniqueName: \"kubernetes.io/projected/ebfe77df-bbcd-412f-9703-78e23d93f410-kube-api-access-8h44k\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.568160 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-public-tls-certs\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.568201 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-combined-ca-bundle\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.571492 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebfe77df-bbcd-412f-9703-78e23d93f410-logs\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.573500 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data-custom\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.577103 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-public-tls-certs\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.577434 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-combined-ca-bundle\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.577705 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.582337 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-internal-tls-certs\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.588549 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h44k\" (UniqueName: \"kubernetes.io/projected/ebfe77df-bbcd-412f-9703-78e23d93f410-kube-api-access-8h44k\") pod \"barbican-api-644b86779b-b4psv\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:06 crc kubenswrapper[4799]: I1124 07:06:06.695688 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:07 crc kubenswrapper[4799]: W1124 07:06:07.256470 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebfe77df_bbcd_412f_9703_78e23d93f410.slice/crio-8bb3ed080c7c54341b172b5767b6c3765958972f0eaabaa31b60e741aa82c318 WatchSource:0}: Error finding container 8bb3ed080c7c54341b172b5767b6c3765958972f0eaabaa31b60e741aa82c318: Status 404 returned error can't find the container with id 8bb3ed080c7c54341b172b5767b6c3765958972f0eaabaa31b60e741aa82c318 Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.256970 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-644b86779b-b4psv"] Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.813037 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644b86779b-b4psv" event={"ID":"ebfe77df-bbcd-412f-9703-78e23d93f410","Type":"ContainerStarted","Data":"2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18"} Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.813334 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644b86779b-b4psv" event={"ID":"ebfe77df-bbcd-412f-9703-78e23d93f410","Type":"ContainerStarted","Data":"c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af"} Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.813349 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644b86779b-b4psv" event={"ID":"ebfe77df-bbcd-412f-9703-78e23d93f410","Type":"ContainerStarted","Data":"8bb3ed080c7c54341b172b5767b6c3765958972f0eaabaa31b60e741aa82c318"} Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.813362 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.815337 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qfbzk" event={"ID":"a987addd-6874-4499-9ea1-27ec674bf28a","Type":"ContainerStarted","Data":"40866b9dd65255e1c5dc3bd4fecb4975eab55bdd3723df4f474e46303e4b4a46"} Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.817385 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" event={"ID":"c986e5b7-9864-46ae-9c99-9072e9c6ce34","Type":"ContainerStarted","Data":"90b897d4e24c66a97b38d28448dae228ba7620bc06b2cd553b7c3edcb5ea7710"} Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.817423 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" event={"ID":"c986e5b7-9864-46ae-9c99-9072e9c6ce34","Type":"ContainerStarted","Data":"d87462cbe65628447244d9ea1c17fb1adb8cfc23702f9f99482cba9b90403d7c"} Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.821838 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7dd777b755-wgf6d" event={"ID":"a9ee9f9f-66be-467a-a982-9ede1e1b682c","Type":"ContainerStarted","Data":"f9a800a0af5f1a9b50d8af38a76b70c4541840dcf2598a2012116de5e32ec921"} Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.821880 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7dd777b755-wgf6d" event={"ID":"a9ee9f9f-66be-467a-a982-9ede1e1b682c","Type":"ContainerStarted","Data":"51856ae48de556e52c655ad669ed192facc4a302f9d9889f4ff2d00441ecb8fa"} Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.843532 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-644b86779b-b4psv" podStartSLOduration=1.843515464 podStartE2EDuration="1.843515464s" podCreationTimestamp="2025-11-24 07:06:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:07.842364152 +0000 UTC m=+1113.498346646" watchObservedRunningTime="2025-11-24 07:06:07.843515464 +0000 UTC m=+1113.499497928" Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.864694 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" podStartSLOduration=2.635131303 podStartE2EDuration="5.864677873s" podCreationTimestamp="2025-11-24 07:06:02 +0000 UTC" firstStartedPulling="2025-11-24 07:06:03.425691993 +0000 UTC m=+1109.081674467" lastFinishedPulling="2025-11-24 07:06:06.655238563 +0000 UTC m=+1112.311221037" observedRunningTime="2025-11-24 07:06:07.861228455 +0000 UTC m=+1113.517210929" watchObservedRunningTime="2025-11-24 07:06:07.864677873 +0000 UTC m=+1113.520660347" Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.893622 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-qfbzk" podStartSLOduration=5.178193326 podStartE2EDuration="43.893604161s" podCreationTimestamp="2025-11-24 07:05:24 +0000 UTC" firstStartedPulling="2025-11-24 07:05:26.36912393 +0000 UTC m=+1072.025106404" lastFinishedPulling="2025-11-24 07:06:05.084534765 +0000 UTC m=+1110.740517239" observedRunningTime="2025-11-24 07:06:07.882052804 +0000 UTC m=+1113.538035278" watchObservedRunningTime="2025-11-24 07:06:07.893604161 +0000 UTC m=+1113.549586635" Nov 24 07:06:07 crc kubenswrapper[4799]: I1124 07:06:07.908095 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7dd777b755-wgf6d" podStartSLOduration=2.573373936 podStartE2EDuration="5.90807538s" podCreationTimestamp="2025-11-24 07:06:02 +0000 UTC" firstStartedPulling="2025-11-24 07:06:03.298883476 +0000 UTC m=+1108.954865950" lastFinishedPulling="2025-11-24 07:06:06.63358492 +0000 UTC m=+1112.289567394" observedRunningTime="2025-11-24 07:06:07.905169538 +0000 UTC m=+1113.561152012" watchObservedRunningTime="2025-11-24 07:06:07.90807538 +0000 UTC m=+1113.564057854" Nov 24 07:06:08 crc kubenswrapper[4799]: I1124 07:06:08.832307 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:11 crc kubenswrapper[4799]: I1124 07:06:11.867276 4799 generic.go:334] "Generic (PLEG): container finished" podID="a987addd-6874-4499-9ea1-27ec674bf28a" containerID="40866b9dd65255e1c5dc3bd4fecb4975eab55bdd3723df4f474e46303e4b4a46" exitCode=0 Nov 24 07:06:11 crc kubenswrapper[4799]: I1124 07:06:11.867360 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qfbzk" event={"ID":"a987addd-6874-4499-9ea1-27ec674bf28a","Type":"ContainerDied","Data":"40866b9dd65255e1c5dc3bd4fecb4975eab55bdd3723df4f474e46303e4b4a46"} Nov 24 07:06:11 crc kubenswrapper[4799]: I1124 07:06:11.970408 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 07:06:11 crc kubenswrapper[4799]: I1124 07:06:11.970519 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.003338 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.003416 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.054786 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.057631 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.059465 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.073183 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.846030 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.882227 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71fffc24-2640-46ca-92aa-f5a2926f5864","Type":"ContainerStarted","Data":"2d076dbd3a37e4a43cd4a34b214cf7f2fdfc93ac9ddfefcfe1068b186f9d4351"} Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.882411 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.882497 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.883281 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="ceilometer-central-agent" containerID="cri-o://bd3eb6eaeec25a6e9fe8c5bb6789ed9b5917e54bffc90e07f048592536b9c6c4" gracePeriod=30 Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.884415 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.884448 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.884523 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="proxy-httpd" containerID="cri-o://2d076dbd3a37e4a43cd4a34b214cf7f2fdfc93ac9ddfefcfe1068b186f9d4351" gracePeriod=30 Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.884597 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="sg-core" containerID="cri-o://239a6f800e16ecbd1a6e367158fc89abbb1733a4c14a5996855ff9fdc8c1b554" gracePeriod=30 Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.884643 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="ceilometer-notification-agent" containerID="cri-o://ed7ae04b94d539a51ace888edb6b2b66f24a3233cebe9b6ec4b0aa1fc5ed9e04" gracePeriod=30 Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.948914 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-94d98"] Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.949216 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" podUID="9774b0e9-c82a-4c51-aa43-251b9220dae3" containerName="dnsmasq-dns" containerID="cri-o://32cdd4ab29d85774618fc3ce4f3b80cf07da822841725b1d95407f40133e58a9" gracePeriod=10 Nov 24 07:06:12 crc kubenswrapper[4799]: I1124 07:06:12.958094 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.5377530569999998 podStartE2EDuration="48.958081945s" podCreationTimestamp="2025-11-24 07:05:24 +0000 UTC" firstStartedPulling="2025-11-24 07:05:26.421243924 +0000 UTC m=+1072.077226398" lastFinishedPulling="2025-11-24 07:06:11.841572772 +0000 UTC m=+1117.497555286" observedRunningTime="2025-11-24 07:06:12.949176893 +0000 UTC m=+1118.605159457" watchObservedRunningTime="2025-11-24 07:06:12.958081945 +0000 UTC m=+1118.614064419" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.455888 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.532358 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-scripts\") pod \"a987addd-6874-4499-9ea1-27ec674bf28a\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.532435 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-combined-ca-bundle\") pod \"a987addd-6874-4499-9ea1-27ec674bf28a\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.532556 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a987addd-6874-4499-9ea1-27ec674bf28a-etc-machine-id\") pod \"a987addd-6874-4499-9ea1-27ec674bf28a\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.532585 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dwj2\" (UniqueName: \"kubernetes.io/projected/a987addd-6874-4499-9ea1-27ec674bf28a-kube-api-access-4dwj2\") pod \"a987addd-6874-4499-9ea1-27ec674bf28a\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.532688 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-config-data\") pod \"a987addd-6874-4499-9ea1-27ec674bf28a\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.532718 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-db-sync-config-data\") pod \"a987addd-6874-4499-9ea1-27ec674bf28a\" (UID: \"a987addd-6874-4499-9ea1-27ec674bf28a\") " Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.534294 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a987addd-6874-4499-9ea1-27ec674bf28a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a987addd-6874-4499-9ea1-27ec674bf28a" (UID: "a987addd-6874-4499-9ea1-27ec674bf28a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.544424 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a987addd-6874-4499-9ea1-27ec674bf28a-kube-api-access-4dwj2" (OuterVolumeSpecName: "kube-api-access-4dwj2") pod "a987addd-6874-4499-9ea1-27ec674bf28a" (UID: "a987addd-6874-4499-9ea1-27ec674bf28a"). InnerVolumeSpecName "kube-api-access-4dwj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.544508 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a987addd-6874-4499-9ea1-27ec674bf28a" (UID: "a987addd-6874-4499-9ea1-27ec674bf28a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.545052 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-scripts" (OuterVolumeSpecName: "scripts") pod "a987addd-6874-4499-9ea1-27ec674bf28a" (UID: "a987addd-6874-4499-9ea1-27ec674bf28a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.594006 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a987addd-6874-4499-9ea1-27ec674bf28a" (UID: "a987addd-6874-4499-9ea1-27ec674bf28a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.635015 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.635041 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.635052 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.635060 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a987addd-6874-4499-9ea1-27ec674bf28a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.635071 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dwj2\" (UniqueName: \"kubernetes.io/projected/a987addd-6874-4499-9ea1-27ec674bf28a-kube-api-access-4dwj2\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.636475 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-config-data" (OuterVolumeSpecName: "config-data") pod "a987addd-6874-4499-9ea1-27ec674bf28a" (UID: "a987addd-6874-4499-9ea1-27ec674bf28a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.738466 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a987addd-6874-4499-9ea1-27ec674bf28a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.891778 4799 generic.go:334] "Generic (PLEG): container finished" podID="9774b0e9-c82a-4c51-aa43-251b9220dae3" containerID="32cdd4ab29d85774618fc3ce4f3b80cf07da822841725b1d95407f40133e58a9" exitCode=0 Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.891878 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" event={"ID":"9774b0e9-c82a-4c51-aa43-251b9220dae3","Type":"ContainerDied","Data":"32cdd4ab29d85774618fc3ce4f3b80cf07da822841725b1d95407f40133e58a9"} Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.894067 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qfbzk" event={"ID":"a987addd-6874-4499-9ea1-27ec674bf28a","Type":"ContainerDied","Data":"6e5ca720cf0da438c6b4e13370ddf0b7b58035de88d9a5a1b82ac52f8c4966c5"} Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.894171 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e5ca720cf0da438c6b4e13370ddf0b7b58035de88d9a5a1b82ac52f8c4966c5" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.894266 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qfbzk" Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.898411 4799 generic.go:334] "Generic (PLEG): container finished" podID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerID="2d076dbd3a37e4a43cd4a34b214cf7f2fdfc93ac9ddfefcfe1068b186f9d4351" exitCode=0 Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.898474 4799 generic.go:334] "Generic (PLEG): container finished" podID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerID="239a6f800e16ecbd1a6e367158fc89abbb1733a4c14a5996855ff9fdc8c1b554" exitCode=2 Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.898484 4799 generic.go:334] "Generic (PLEG): container finished" podID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerID="bd3eb6eaeec25a6e9fe8c5bb6789ed9b5917e54bffc90e07f048592536b9c6c4" exitCode=0 Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.898768 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71fffc24-2640-46ca-92aa-f5a2926f5864","Type":"ContainerDied","Data":"2d076dbd3a37e4a43cd4a34b214cf7f2fdfc93ac9ddfefcfe1068b186f9d4351"} Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.898874 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71fffc24-2640-46ca-92aa-f5a2926f5864","Type":"ContainerDied","Data":"239a6f800e16ecbd1a6e367158fc89abbb1733a4c14a5996855ff9fdc8c1b554"} Nov 24 07:06:13 crc kubenswrapper[4799]: I1124 07:06:13.898965 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71fffc24-2640-46ca-92aa-f5a2926f5864","Type":"ContainerDied","Data":"bd3eb6eaeec25a6e9fe8c5bb6789ed9b5917e54bffc90e07f048592536b9c6c4"} Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.289448 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-797bbc649-tc48z"] Nov 24 07:06:14 crc kubenswrapper[4799]: E1124 07:06:14.290320 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a987addd-6874-4499-9ea1-27ec674bf28a" containerName="cinder-db-sync" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.290340 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a987addd-6874-4499-9ea1-27ec674bf28a" containerName="cinder-db-sync" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.290551 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a987addd-6874-4499-9ea1-27ec674bf28a" containerName="cinder-db-sync" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.291721 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.321997 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-tc48z"] Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.347914 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d2gk\" (UniqueName: \"kubernetes.io/projected/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-kube-api-access-6d2gk\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.347970 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-nb\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.347995 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-sb\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.348019 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-config\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.348078 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-svc\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.348138 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-swift-storage-0\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.414955 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.416667 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.453492 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.453892 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.454200 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-tsmnv" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.460543 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.461654 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.498306 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-swift-storage-0\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.498585 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.498697 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.498839 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-scripts\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.498983 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d2gk\" (UniqueName: \"kubernetes.io/projected/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-kube-api-access-6d2gk\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.499054 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-nb\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.499128 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-sb\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.499233 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-config\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.499352 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-svc\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.499442 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.499515 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrjzj\" (UniqueName: \"kubernetes.io/projected/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-kube-api-access-qrjzj\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.499565 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-swift-storage-0\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.499647 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.500492 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-svc\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.500815 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-nb\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.500831 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-sb\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.507729 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-config\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.561976 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d2gk\" (UniqueName: \"kubernetes.io/projected/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-kube-api-access-6d2gk\") pod \"dnsmasq-dns-797bbc649-tc48z\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.594602 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.597787 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.602774 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.602823 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrjzj\" (UniqueName: \"kubernetes.io/projected/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-kube-api-access-qrjzj\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.602859 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.602913 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.602932 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.602993 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-scripts\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.608001 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.613162 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.622738 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.626777 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.627360 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.636621 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrjzj\" (UniqueName: \"kubernetes.io/projected/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-kube-api-access-qrjzj\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.638917 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.639179 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-scripts\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.663555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.706995 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwlpj\" (UniqueName: \"kubernetes.io/projected/69f24266-9596-4aa0-ba91-1b838b0be93a-kube-api-access-zwlpj\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.707059 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f24266-9596-4aa0-ba91-1b838b0be93a-logs\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.707086 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data-custom\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.707110 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/69f24266-9596-4aa0-ba91-1b838b0be93a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.707125 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.707146 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.707191 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-scripts\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.815901 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-scripts\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.816127 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwlpj\" (UniqueName: \"kubernetes.io/projected/69f24266-9596-4aa0-ba91-1b838b0be93a-kube-api-access-zwlpj\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.816204 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f24266-9596-4aa0-ba91-1b838b0be93a-logs\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.816264 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data-custom\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.816293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/69f24266-9596-4aa0-ba91-1b838b0be93a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.816310 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.816355 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.818327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f24266-9596-4aa0-ba91-1b838b0be93a-logs\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.818980 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/69f24266-9596-4aa0-ba91-1b838b0be93a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.827430 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data-custom\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.827555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.830406 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.835781 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-scripts\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.854516 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.889653 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwlpj\" (UniqueName: \"kubernetes.io/projected/69f24266-9596-4aa0-ba91-1b838b0be93a-kube-api-access-zwlpj\") pod \"cinder-api-0\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.913559 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.964665 4799 generic.go:334] "Generic (PLEG): container finished" podID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerID="ed7ae04b94d539a51ace888edb6b2b66f24a3233cebe9b6ec4b0aa1fc5ed9e04" exitCode=0 Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.964824 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.964833 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.966324 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71fffc24-2640-46ca-92aa-f5a2926f5864","Type":"ContainerDied","Data":"ed7ae04b94d539a51ace888edb6b2b66f24a3233cebe9b6ec4b0aa1fc5ed9e04"} Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.966568 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 07:06:14 crc kubenswrapper[4799]: I1124 07:06:14.966634 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.020784 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.136506 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-dns-svc\") pod \"9774b0e9-c82a-4c51-aa43-251b9220dae3\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.136621 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6lj4\" (UniqueName: \"kubernetes.io/projected/9774b0e9-c82a-4c51-aa43-251b9220dae3-kube-api-access-b6lj4\") pod \"9774b0e9-c82a-4c51-aa43-251b9220dae3\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.136650 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-sb\") pod \"9774b0e9-c82a-4c51-aa43-251b9220dae3\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.136740 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-config\") pod \"9774b0e9-c82a-4c51-aa43-251b9220dae3\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.136783 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-nb\") pod \"9774b0e9-c82a-4c51-aa43-251b9220dae3\" (UID: \"9774b0e9-c82a-4c51-aa43-251b9220dae3\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.158174 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9774b0e9-c82a-4c51-aa43-251b9220dae3-kube-api-access-b6lj4" (OuterVolumeSpecName: "kube-api-access-b6lj4") pod "9774b0e9-c82a-4c51-aa43-251b9220dae3" (UID: "9774b0e9-c82a-4c51-aa43-251b9220dae3"). InnerVolumeSpecName "kube-api-access-b6lj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.242674 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6lj4\" (UniqueName: \"kubernetes.io/projected/9774b0e9-c82a-4c51-aa43-251b9220dae3-kube-api-access-b6lj4\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.245601 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9774b0e9-c82a-4c51-aa43-251b9220dae3" (UID: "9774b0e9-c82a-4c51-aa43-251b9220dae3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.308617 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9774b0e9-c82a-4c51-aa43-251b9220dae3" (UID: "9774b0e9-c82a-4c51-aa43-251b9220dae3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.345914 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.345960 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.376436 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9774b0e9-c82a-4c51-aa43-251b9220dae3" (UID: "9774b0e9-c82a-4c51-aa43-251b9220dae3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.425634 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-config" (OuterVolumeSpecName: "config") pod "9774b0e9-c82a-4c51-aa43-251b9220dae3" (UID: "9774b0e9-c82a-4c51-aa43-251b9220dae3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.448635 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.448996 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9774b0e9-c82a-4c51-aa43-251b9220dae3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.741202 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-tc48z"] Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.806786 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.867880 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.875412 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-log-httpd\") pod \"71fffc24-2640-46ca-92aa-f5a2926f5864\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.875523 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-combined-ca-bundle\") pod \"71fffc24-2640-46ca-92aa-f5a2926f5864\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.875574 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-config-data\") pod \"71fffc24-2640-46ca-92aa-f5a2926f5864\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.875595 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjj5s\" (UniqueName: \"kubernetes.io/projected/71fffc24-2640-46ca-92aa-f5a2926f5864-kube-api-access-zjj5s\") pod \"71fffc24-2640-46ca-92aa-f5a2926f5864\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.875682 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-sg-core-conf-yaml\") pod \"71fffc24-2640-46ca-92aa-f5a2926f5864\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.875703 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-scripts\") pod \"71fffc24-2640-46ca-92aa-f5a2926f5864\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.875736 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-run-httpd\") pod \"71fffc24-2640-46ca-92aa-f5a2926f5864\" (UID: \"71fffc24-2640-46ca-92aa-f5a2926f5864\") " Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.876368 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "71fffc24-2640-46ca-92aa-f5a2926f5864" (UID: "71fffc24-2640-46ca-92aa-f5a2926f5864"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.878986 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "71fffc24-2640-46ca-92aa-f5a2926f5864" (UID: "71fffc24-2640-46ca-92aa-f5a2926f5864"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.894158 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-scripts" (OuterVolumeSpecName: "scripts") pod "71fffc24-2640-46ca-92aa-f5a2926f5864" (UID: "71fffc24-2640-46ca-92aa-f5a2926f5864"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.914693 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71fffc24-2640-46ca-92aa-f5a2926f5864-kube-api-access-zjj5s" (OuterVolumeSpecName: "kube-api-access-zjj5s") pod "71fffc24-2640-46ca-92aa-f5a2926f5864" (UID: "71fffc24-2640-46ca-92aa-f5a2926f5864"). InnerVolumeSpecName "kube-api-access-zjj5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.945130 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "71fffc24-2640-46ca-92aa-f5a2926f5864" (UID: "71fffc24-2640-46ca-92aa-f5a2926f5864"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.947422 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.978148 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.978454 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.978464 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.978471 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71fffc24-2640-46ca-92aa-f5a2926f5864-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:15 crc kubenswrapper[4799]: I1124 07:06:15.978480 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjj5s\" (UniqueName: \"kubernetes.io/projected/71fffc24-2640-46ca-92aa-f5a2926f5864-kube-api-access-zjj5s\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.041600 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71fffc24-2640-46ca-92aa-f5a2926f5864" (UID: "71fffc24-2640-46ca-92aa-f5a2926f5864"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.043543 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71fffc24-2640-46ca-92aa-f5a2926f5864","Type":"ContainerDied","Data":"b0ed93876a566b5169ba2512b5b3cc1b3c5678feba1004428c91b4b189a351ed"} Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.043606 4799 scope.go:117] "RemoveContainer" containerID="2d076dbd3a37e4a43cd4a34b214cf7f2fdfc93ac9ddfefcfe1068b186f9d4351" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.043764 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.064025 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-tc48z" event={"ID":"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38","Type":"ContainerStarted","Data":"19d0ad5d6052c714e6dbf14a19cb13e64d58b305d7d17f1214a9067f801e7af9"} Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.067737 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.074191 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" event={"ID":"9774b0e9-c82a-4c51-aa43-251b9220dae3","Type":"ContainerDied","Data":"f38b3ff57bf7c89677d8186f97f02dae73bc1814f026471b8fec822a1cb4e1ab"} Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.074342 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-94d98" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.080837 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.081167 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-config-data" (OuterVolumeSpecName: "config-data") pod "71fffc24-2640-46ca-92aa-f5a2926f5864" (UID: "71fffc24-2640-46ca-92aa-f5a2926f5864"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.101202 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab","Type":"ContainerStarted","Data":"383c79db7baa1d204c0f2cff05798b094926b21607cfc4c4c9aefa6edca3e7aa"} Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.183770 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71fffc24-2640-46ca-92aa-f5a2926f5864-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.205345 4799 scope.go:117] "RemoveContainer" containerID="239a6f800e16ecbd1a6e367158fc89abbb1733a4c14a5996855ff9fdc8c1b554" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.233899 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-94d98"] Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.247199 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-94d98"] Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.282997 4799 scope.go:117] "RemoveContainer" containerID="ed7ae04b94d539a51ace888edb6b2b66f24a3233cebe9b6ec4b0aa1fc5ed9e04" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.383685 4799 scope.go:117] "RemoveContainer" containerID="bd3eb6eaeec25a6e9fe8c5bb6789ed9b5917e54bffc90e07f048592536b9c6c4" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.410058 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.421148 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.464901 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:16 crc kubenswrapper[4799]: E1124 07:06:16.465339 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="sg-core" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.465352 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="sg-core" Nov 24 07:06:16 crc kubenswrapper[4799]: E1124 07:06:16.465395 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9774b0e9-c82a-4c51-aa43-251b9220dae3" containerName="init" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.465402 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9774b0e9-c82a-4c51-aa43-251b9220dae3" containerName="init" Nov 24 07:06:16 crc kubenswrapper[4799]: E1124 07:06:16.465422 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="proxy-httpd" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.465443 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="proxy-httpd" Nov 24 07:06:16 crc kubenswrapper[4799]: E1124 07:06:16.465464 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="ceilometer-central-agent" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.465471 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="ceilometer-central-agent" Nov 24 07:06:16 crc kubenswrapper[4799]: E1124 07:06:16.465481 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="ceilometer-notification-agent" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.465490 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="ceilometer-notification-agent" Nov 24 07:06:16 crc kubenswrapper[4799]: E1124 07:06:16.465511 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9774b0e9-c82a-4c51-aa43-251b9220dae3" containerName="dnsmasq-dns" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.465519 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9774b0e9-c82a-4c51-aa43-251b9220dae3" containerName="dnsmasq-dns" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.465717 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="proxy-httpd" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.465751 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9774b0e9-c82a-4c51-aa43-251b9220dae3" containerName="dnsmasq-dns" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.465762 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="sg-core" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.465773 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="ceilometer-central-agent" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.465784 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" containerName="ceilometer-notification-agent" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.467338 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.470675 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.472828 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.489658 4799 scope.go:117] "RemoveContainer" containerID="32cdd4ab29d85774618fc3ce4f3b80cf07da822841725b1d95407f40133e58a9" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.507734 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.507797 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw9fx\" (UniqueName: \"kubernetes.io/projected/d4c442bc-d661-4294-a156-a0ad37eebc7c-kube-api-access-fw9fx\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.507861 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-scripts\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.507903 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.507990 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-run-httpd\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.508017 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-log-httpd\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.508054 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-config-data\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.566974 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.570136 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.615311 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-log-httpd\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.615370 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-config-data\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.615447 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw9fx\" (UniqueName: \"kubernetes.io/projected/d4c442bc-d661-4294-a156-a0ad37eebc7c-kube-api-access-fw9fx\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.615537 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.615589 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-scripts\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.615628 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.615676 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-run-httpd\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.617512 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-log-httpd\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.618135 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-run-httpd\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.631031 4799 scope.go:117] "RemoveContainer" containerID="847abe5903aab2744f125f425787c7b139e9c7c11ce8a965535b10689fc42575" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.642899 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.643965 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-config-data\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.644628 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.645811 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-scripts\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.668785 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw9fx\" (UniqueName: \"kubernetes.io/projected/d4c442bc-d661-4294-a156-a0ad37eebc7c-kube-api-access-fw9fx\") pod \"ceilometer-0\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.812197 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.812319 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.846522 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:06:16 crc kubenswrapper[4799]: I1124 07:06:16.995395 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:06:17 crc kubenswrapper[4799]: I1124 07:06:17.196894 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"69f24266-9596-4aa0-ba91-1b838b0be93a","Type":"ContainerStarted","Data":"2e82469818c252fc3aba4328cd46702a3f278449f12eb949956a1085a232bc95"} Nov 24 07:06:17 crc kubenswrapper[4799]: I1124 07:06:17.212047 4799 generic.go:334] "Generic (PLEG): container finished" podID="f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" containerID="589b81b002f0c3ac98becae0625aff8bf97cbfa6259f033886a8fe5e8e3f5618" exitCode=0 Nov 24 07:06:17 crc kubenswrapper[4799]: I1124 07:06:17.212124 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-tc48z" event={"ID":"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38","Type":"ContainerDied","Data":"589b81b002f0c3ac98becae0625aff8bf97cbfa6259f033886a8fe5e8e3f5618"} Nov 24 07:06:17 crc kubenswrapper[4799]: I1124 07:06:17.264022 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 07:06:17 crc kubenswrapper[4799]: I1124 07:06:17.264172 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 07:06:17 crc kubenswrapper[4799]: I1124 07:06:17.323325 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 07:06:17 crc kubenswrapper[4799]: I1124 07:06:17.514810 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 07:06:17 crc kubenswrapper[4799]: I1124 07:06:17.660552 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71fffc24-2640-46ca-92aa-f5a2926f5864" path="/var/lib/kubelet/pods/71fffc24-2640-46ca-92aa-f5a2926f5864/volumes" Nov 24 07:06:17 crc kubenswrapper[4799]: I1124 07:06:17.662532 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9774b0e9-c82a-4c51-aa43-251b9220dae3" path="/var/lib/kubelet/pods/9774b0e9-c82a-4c51-aa43-251b9220dae3/volumes" Nov 24 07:06:17 crc kubenswrapper[4799]: I1124 07:06:17.895192 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:17 crc kubenswrapper[4799]: W1124 07:06:17.914822 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4c442bc_d661_4294_a156_a0ad37eebc7c.slice/crio-c90f92146c1825de25aa9c53ca4f215be6b89d17999a5a712d9db4048c58eed1 WatchSource:0}: Error finding container c90f92146c1825de25aa9c53ca4f215be6b89d17999a5a712d9db4048c58eed1: Status 404 returned error can't find the container with id c90f92146c1825de25aa9c53ca4f215be6b89d17999a5a712d9db4048c58eed1 Nov 24 07:06:18 crc kubenswrapper[4799]: I1124 07:06:18.230469 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"69f24266-9596-4aa0-ba91-1b838b0be93a","Type":"ContainerStarted","Data":"c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb"} Nov 24 07:06:18 crc kubenswrapper[4799]: I1124 07:06:18.241672 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4c442bc-d661-4294-a156-a0ad37eebc7c","Type":"ContainerStarted","Data":"c90f92146c1825de25aa9c53ca4f215be6b89d17999a5a712d9db4048c58eed1"} Nov 24 07:06:18 crc kubenswrapper[4799]: I1124 07:06:18.250807 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-tc48z" event={"ID":"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38","Type":"ContainerStarted","Data":"50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533"} Nov 24 07:06:18 crc kubenswrapper[4799]: I1124 07:06:18.250892 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:18 crc kubenswrapper[4799]: I1124 07:06:18.277500 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-797bbc649-tc48z" podStartSLOduration=4.277463817 podStartE2EDuration="4.277463817s" podCreationTimestamp="2025-11-24 07:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:18.271257772 +0000 UTC m=+1123.927240246" watchObservedRunningTime="2025-11-24 07:06:18.277463817 +0000 UTC m=+1123.933446291" Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.270400 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab","Type":"ContainerStarted","Data":"391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f"} Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.287940 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4c442bc-d661-4294-a156-a0ad37eebc7c","Type":"ContainerStarted","Data":"77e3d849b1ab7a53dd8b46c090efd860e8595b4933e9efb477e2ce9bac051e9c"} Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.296616 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.299675 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="69f24266-9596-4aa0-ba91-1b838b0be93a" containerName="cinder-api-log" containerID="cri-o://c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb" gracePeriod=30 Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.300061 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"69f24266-9596-4aa0-ba91-1b838b0be93a","Type":"ContainerStarted","Data":"d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e"} Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.300114 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.300147 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="69f24266-9596-4aa0-ba91-1b838b0be93a" containerName="cinder-api" containerID="cri-o://d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e" gracePeriod=30 Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.340287 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.34027072 podStartE2EDuration="5.34027072s" podCreationTimestamp="2025-11-24 07:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:19.336804982 +0000 UTC m=+1124.992787456" watchObservedRunningTime="2025-11-24 07:06:19.34027072 +0000 UTC m=+1124.996253194" Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.403273 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.499721 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d7959f94d-skwn6"] Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.522277 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d7959f94d-skwn6" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api-log" containerID="cri-o://f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691" gracePeriod=30 Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.522683 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d7959f94d-skwn6" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api" containerID="cri-o://4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d" gracePeriod=30 Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.536821 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5d7959f94d-skwn6" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Nov 24 07:06:19 crc kubenswrapper[4799]: I1124 07:06:19.536998 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7959f94d-skwn6" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Nov 24 07:06:20 crc kubenswrapper[4799]: I1124 07:06:20.321127 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab","Type":"ContainerStarted","Data":"aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27"} Nov 24 07:06:20 crc kubenswrapper[4799]: I1124 07:06:20.325868 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4c442bc-d661-4294-a156-a0ad37eebc7c","Type":"ContainerStarted","Data":"3572b6c411d077e12bdd5010147fb9798945058a7833556cd2f07866a4bbe975"} Nov 24 07:06:20 crc kubenswrapper[4799]: I1124 07:06:20.332482 4799 generic.go:334] "Generic (PLEG): container finished" podID="f633fe14-fd8c-4738-9960-4fff9619c027" containerID="f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691" exitCode=143 Nov 24 07:06:20 crc kubenswrapper[4799]: I1124 07:06:20.332527 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7959f94d-skwn6" event={"ID":"f633fe14-fd8c-4738-9960-4fff9619c027","Type":"ContainerDied","Data":"f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691"} Nov 24 07:06:20 crc kubenswrapper[4799]: I1124 07:06:20.342395 4799 generic.go:334] "Generic (PLEG): container finished" podID="69f24266-9596-4aa0-ba91-1b838b0be93a" containerID="c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb" exitCode=143 Nov 24 07:06:20 crc kubenswrapper[4799]: I1124 07:06:20.343721 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"69f24266-9596-4aa0-ba91-1b838b0be93a","Type":"ContainerDied","Data":"c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb"} Nov 24 07:06:20 crc kubenswrapper[4799]: I1124 07:06:20.354887 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.01568067 podStartE2EDuration="6.354865909s" podCreationTimestamp="2025-11-24 07:06:14 +0000 UTC" firstStartedPulling="2025-11-24 07:06:15.915803656 +0000 UTC m=+1121.571786130" lastFinishedPulling="2025-11-24 07:06:17.254988895 +0000 UTC m=+1122.910971369" observedRunningTime="2025-11-24 07:06:20.34467228 +0000 UTC m=+1126.000654764" watchObservedRunningTime="2025-11-24 07:06:20.354865909 +0000 UTC m=+1126.010848463" Nov 24 07:06:20 crc kubenswrapper[4799]: I1124 07:06:20.400464 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:06:20 crc kubenswrapper[4799]: I1124 07:06:20.400531 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.201766 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.252439 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f24266-9596-4aa0-ba91-1b838b0be93a-logs\") pod \"69f24266-9596-4aa0-ba91-1b838b0be93a\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.252504 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/69f24266-9596-4aa0-ba91-1b838b0be93a-etc-machine-id\") pod \"69f24266-9596-4aa0-ba91-1b838b0be93a\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.252537 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-scripts\") pod \"69f24266-9596-4aa0-ba91-1b838b0be93a\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.252628 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data-custom\") pod \"69f24266-9596-4aa0-ba91-1b838b0be93a\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.252654 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwlpj\" (UniqueName: \"kubernetes.io/projected/69f24266-9596-4aa0-ba91-1b838b0be93a-kube-api-access-zwlpj\") pod \"69f24266-9596-4aa0-ba91-1b838b0be93a\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.252731 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-combined-ca-bundle\") pod \"69f24266-9596-4aa0-ba91-1b838b0be93a\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.252754 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data\") pod \"69f24266-9596-4aa0-ba91-1b838b0be93a\" (UID: \"69f24266-9596-4aa0-ba91-1b838b0be93a\") " Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.255881 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69f24266-9596-4aa0-ba91-1b838b0be93a-logs" (OuterVolumeSpecName: "logs") pod "69f24266-9596-4aa0-ba91-1b838b0be93a" (UID: "69f24266-9596-4aa0-ba91-1b838b0be93a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.255939 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69f24266-9596-4aa0-ba91-1b838b0be93a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "69f24266-9596-4aa0-ba91-1b838b0be93a" (UID: "69f24266-9596-4aa0-ba91-1b838b0be93a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.261460 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-scripts" (OuterVolumeSpecName: "scripts") pod "69f24266-9596-4aa0-ba91-1b838b0be93a" (UID: "69f24266-9596-4aa0-ba91-1b838b0be93a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.262674 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69f24266-9596-4aa0-ba91-1b838b0be93a-kube-api-access-zwlpj" (OuterVolumeSpecName: "kube-api-access-zwlpj") pod "69f24266-9596-4aa0-ba91-1b838b0be93a" (UID: "69f24266-9596-4aa0-ba91-1b838b0be93a"). InnerVolumeSpecName "kube-api-access-zwlpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.280061 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "69f24266-9596-4aa0-ba91-1b838b0be93a" (UID: "69f24266-9596-4aa0-ba91-1b838b0be93a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.284816 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69f24266-9596-4aa0-ba91-1b838b0be93a" (UID: "69f24266-9596-4aa0-ba91-1b838b0be93a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.340883 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data" (OuterVolumeSpecName: "config-data") pod "69f24266-9596-4aa0-ba91-1b838b0be93a" (UID: "69f24266-9596-4aa0-ba91-1b838b0be93a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.355688 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f24266-9596-4aa0-ba91-1b838b0be93a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.355740 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/69f24266-9596-4aa0-ba91-1b838b0be93a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.355751 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.355765 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.355783 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwlpj\" (UniqueName: \"kubernetes.io/projected/69f24266-9596-4aa0-ba91-1b838b0be93a-kube-api-access-zwlpj\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.355795 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.355811 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69f24266-9596-4aa0-ba91-1b838b0be93a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.367768 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4c442bc-d661-4294-a156-a0ad37eebc7c","Type":"ContainerStarted","Data":"2e90088664ca828c104ba6d3f4557da6a5523afb97eb2c9138f9a5e9f356b782"} Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.373397 4799 generic.go:334] "Generic (PLEG): container finished" podID="69f24266-9596-4aa0-ba91-1b838b0be93a" containerID="d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e" exitCode=0 Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.374152 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"69f24266-9596-4aa0-ba91-1b838b0be93a","Type":"ContainerDied","Data":"d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e"} Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.374181 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.374219 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"69f24266-9596-4aa0-ba91-1b838b0be93a","Type":"ContainerDied","Data":"2e82469818c252fc3aba4328cd46702a3f278449f12eb949956a1085a232bc95"} Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.374244 4799 scope.go:117] "RemoveContainer" containerID="d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.414785 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.421337 4799 scope.go:117] "RemoveContainer" containerID="c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.433626 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.452316 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:06:21 crc kubenswrapper[4799]: E1124 07:06:21.452862 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f24266-9596-4aa0-ba91-1b838b0be93a" containerName="cinder-api" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.452947 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f24266-9596-4aa0-ba91-1b838b0be93a" containerName="cinder-api" Nov 24 07:06:21 crc kubenswrapper[4799]: E1124 07:06:21.453037 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f24266-9596-4aa0-ba91-1b838b0be93a" containerName="cinder-api-log" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.453150 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f24266-9596-4aa0-ba91-1b838b0be93a" containerName="cinder-api-log" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.453364 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="69f24266-9596-4aa0-ba91-1b838b0be93a" containerName="cinder-api" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.453438 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="69f24266-9596-4aa0-ba91-1b838b0be93a" containerName="cinder-api-log" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.454446 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.458016 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f643414c-72f0-4e7c-b804-20a56f3207ab-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.458058 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dk9x\" (UniqueName: \"kubernetes.io/projected/f643414c-72f0-4e7c-b804-20a56f3207ab-kube-api-access-9dk9x\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.458159 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.458179 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f643414c-72f0-4e7c-b804-20a56f3207ab-logs\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.458219 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-scripts\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.458241 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.458255 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.458284 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.458328 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data-custom\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.460831 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.461093 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.461317 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.472350 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.489009 4799 scope.go:117] "RemoveContainer" containerID="d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e" Nov 24 07:06:21 crc kubenswrapper[4799]: E1124 07:06:21.494963 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e\": container with ID starting with d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e not found: ID does not exist" containerID="d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.495019 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e"} err="failed to get container status \"d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e\": rpc error: code = NotFound desc = could not find container \"d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e\": container with ID starting with d257f87d24f6d6628a8a704d2ba14e5f1bbecf7a883bfc63f84fafcf4c12957e not found: ID does not exist" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.495048 4799 scope.go:117] "RemoveContainer" containerID="c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb" Nov 24 07:06:21 crc kubenswrapper[4799]: E1124 07:06:21.496705 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb\": container with ID starting with c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb not found: ID does not exist" containerID="c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.496734 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb"} err="failed to get container status \"c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb\": rpc error: code = NotFound desc = could not find container \"c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb\": container with ID starting with c92139c53d7841f015a95d1c41be821251747e0c2be632663ac3ef493cb19dcb not found: ID does not exist" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.561124 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-scripts\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.561192 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.561223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.561274 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.561306 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data-custom\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.561359 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f643414c-72f0-4e7c-b804-20a56f3207ab-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.561403 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dk9x\" (UniqueName: \"kubernetes.io/projected/f643414c-72f0-4e7c-b804-20a56f3207ab-kube-api-access-9dk9x\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.561467 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f643414c-72f0-4e7c-b804-20a56f3207ab-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.561557 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.561593 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f643414c-72f0-4e7c-b804-20a56f3207ab-logs\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.562763 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f643414c-72f0-4e7c-b804-20a56f3207ab-logs\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.567534 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.567735 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.569348 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-scripts\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.569772 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.570335 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data-custom\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.572445 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.582044 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dk9x\" (UniqueName: \"kubernetes.io/projected/f643414c-72f0-4e7c-b804-20a56f3207ab-kube-api-access-9dk9x\") pod \"cinder-api-0\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " pod="openstack/cinder-api-0" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.644466 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69f24266-9596-4aa0-ba91-1b838b0be93a" path="/var/lib/kubelet/pods/69f24266-9596-4aa0-ba91-1b838b0be93a/volumes" Nov 24 07:06:21 crc kubenswrapper[4799]: I1124 07:06:21.777499 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:06:22 crc kubenswrapper[4799]: I1124 07:06:22.278203 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:06:22 crc kubenswrapper[4799]: W1124 07:06:22.288947 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf643414c_72f0_4e7c_b804_20a56f3207ab.slice/crio-6132ba1992328a5bbcc5e80d2617c0eda719326e31d71528f2970cfeb453f488 WatchSource:0}: Error finding container 6132ba1992328a5bbcc5e80d2617c0eda719326e31d71528f2970cfeb453f488: Status 404 returned error can't find the container with id 6132ba1992328a5bbcc5e80d2617c0eda719326e31d71528f2970cfeb453f488 Nov 24 07:06:22 crc kubenswrapper[4799]: I1124 07:06:22.415210 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f643414c-72f0-4e7c-b804-20a56f3207ab","Type":"ContainerStarted","Data":"6132ba1992328a5bbcc5e80d2617c0eda719326e31d71528f2970cfeb453f488"} Nov 24 07:06:22 crc kubenswrapper[4799]: I1124 07:06:22.424209 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4c442bc-d661-4294-a156-a0ad37eebc7c","Type":"ContainerStarted","Data":"a79bf0ce1dbcde8595fb25f6ca326209543f8a872e833afc9b35f105f1ec00f4"} Nov 24 07:06:22 crc kubenswrapper[4799]: I1124 07:06:22.424534 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:06:22 crc kubenswrapper[4799]: I1124 07:06:22.452155 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7390009920000002 podStartE2EDuration="6.452129721s" podCreationTimestamp="2025-11-24 07:06:16 +0000 UTC" firstStartedPulling="2025-11-24 07:06:17.92508543 +0000 UTC m=+1123.581067904" lastFinishedPulling="2025-11-24 07:06:21.638214159 +0000 UTC m=+1127.294196633" observedRunningTime="2025-11-24 07:06:22.451771131 +0000 UTC m=+1128.107753615" watchObservedRunningTime="2025-11-24 07:06:22.452129721 +0000 UTC m=+1128.108112195" Nov 24 07:06:23 crc kubenswrapper[4799]: I1124 07:06:23.454914 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f643414c-72f0-4e7c-b804-20a56f3207ab","Type":"ContainerStarted","Data":"1bd8daa785503f4e2fcb67f8d5e0fa713f93ffd5b803f47263907f5420e39c50"} Nov 24 07:06:23 crc kubenswrapper[4799]: I1124 07:06:23.971690 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7959f94d-skwn6" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:47458->10.217.0.160:9311: read: connection reset by peer" Nov 24 07:06:23 crc kubenswrapper[4799]: I1124 07:06:23.972967 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7959f94d-skwn6" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:40726->10.217.0.160:9311: read: connection reset by peer" Nov 24 07:06:23 crc kubenswrapper[4799]: I1124 07:06:23.973051 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7959f94d-skwn6" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": dial tcp 10.217.0.160:9311: connect: connection refused" Nov 24 07:06:23 crc kubenswrapper[4799]: I1124 07:06:23.973217 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.443069 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.465748 4799 generic.go:334] "Generic (PLEG): container finished" podID="f633fe14-fd8c-4738-9960-4fff9619c027" containerID="4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d" exitCode=0 Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.465820 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7959f94d-skwn6" event={"ID":"f633fe14-fd8c-4738-9960-4fff9619c027","Type":"ContainerDied","Data":"4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d"} Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.465872 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7959f94d-skwn6" event={"ID":"f633fe14-fd8c-4738-9960-4fff9619c027","Type":"ContainerDied","Data":"b69ba6812367bc13da1033eb9600fd77dbf01dc68bbbc1ee747a04e869827152"} Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.465896 4799 scope.go:117] "RemoveContainer" containerID="4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.466034 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7959f94d-skwn6" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.480968 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f643414c-72f0-4e7c-b804-20a56f3207ab","Type":"ContainerStarted","Data":"30959909b72adeb936fc238d713e097048621efd99d9e5661c63d6da5128b9e8"} Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.481262 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.506988 4799 scope.go:117] "RemoveContainer" containerID="f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.511288 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.511268385 podStartE2EDuration="3.511268385s" podCreationTimestamp="2025-11-24 07:06:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:24.508176937 +0000 UTC m=+1130.164159411" watchObservedRunningTime="2025-11-24 07:06:24.511268385 +0000 UTC m=+1130.167250859" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.537781 4799 scope.go:117] "RemoveContainer" containerID="4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d" Nov 24 07:06:24 crc kubenswrapper[4799]: E1124 07:06:24.538274 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d\": container with ID starting with 4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d not found: ID does not exist" containerID="4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.538321 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d"} err="failed to get container status \"4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d\": rpc error: code = NotFound desc = could not find container \"4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d\": container with ID starting with 4156f033a6b55e2219057184a087f479d32cf70c863d803c8e1deaced75d436d not found: ID does not exist" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.538347 4799 scope.go:117] "RemoveContainer" containerID="f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691" Nov 24 07:06:24 crc kubenswrapper[4799]: E1124 07:06:24.538720 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691\": container with ID starting with f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691 not found: ID does not exist" containerID="f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.538772 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691"} err="failed to get container status \"f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691\": rpc error: code = NotFound desc = could not find container \"f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691\": container with ID starting with f17ad02f5e9f63fdec8b4566b9cb261eac9c4306bb1431dc5ea1b009eebc3691 not found: ID does not exist" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.629186 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.633371 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data\") pod \"f633fe14-fd8c-4738-9960-4fff9619c027\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.633432 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f633fe14-fd8c-4738-9960-4fff9619c027-logs\") pod \"f633fe14-fd8c-4738-9960-4fff9619c027\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.633491 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5v2x9\" (UniqueName: \"kubernetes.io/projected/f633fe14-fd8c-4738-9960-4fff9619c027-kube-api-access-5v2x9\") pod \"f633fe14-fd8c-4738-9960-4fff9619c027\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.633523 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-combined-ca-bundle\") pod \"f633fe14-fd8c-4738-9960-4fff9619c027\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.633544 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data-custom\") pod \"f633fe14-fd8c-4738-9960-4fff9619c027\" (UID: \"f633fe14-fd8c-4738-9960-4fff9619c027\") " Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.634349 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f633fe14-fd8c-4738-9960-4fff9619c027-logs" (OuterVolumeSpecName: "logs") pod "f633fe14-fd8c-4738-9960-4fff9619c027" (UID: "f633fe14-fd8c-4738-9960-4fff9619c027"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.642256 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f633fe14-fd8c-4738-9960-4fff9619c027-kube-api-access-5v2x9" (OuterVolumeSpecName: "kube-api-access-5v2x9") pod "f633fe14-fd8c-4738-9960-4fff9619c027" (UID: "f633fe14-fd8c-4738-9960-4fff9619c027"). InnerVolumeSpecName "kube-api-access-5v2x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.642422 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f633fe14-fd8c-4738-9960-4fff9619c027" (UID: "f633fe14-fd8c-4738-9960-4fff9619c027"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.689118 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f633fe14-fd8c-4738-9960-4fff9619c027" (UID: "f633fe14-fd8c-4738-9960-4fff9619c027"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.714403 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data" (OuterVolumeSpecName: "config-data") pod "f633fe14-fd8c-4738-9960-4fff9619c027" (UID: "f633fe14-fd8c-4738-9960-4fff9619c027"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.720377 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-nwn2f"] Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.720728 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" podUID="3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" containerName="dnsmasq-dns" containerID="cri-o://3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb" gracePeriod=10 Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.739528 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.740197 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f633fe14-fd8c-4738-9960-4fff9619c027-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.740224 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5v2x9\" (UniqueName: \"kubernetes.io/projected/f633fe14-fd8c-4738-9960-4fff9619c027-kube-api-access-5v2x9\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.740239 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.740250 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f633fe14-fd8c-4738-9960-4fff9619c027-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.856000 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.859221 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d7959f94d-skwn6"] Nov 24 07:06:24 crc kubenswrapper[4799]: I1124 07:06:24.868511 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5d7959f94d-skwn6"] Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.194958 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.325838 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.357045 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-swift-storage-0\") pod \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.357122 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-sb\") pod \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.357197 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-config\") pod \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.357223 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-nb\") pod \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.357266 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh9ks\" (UniqueName: \"kubernetes.io/projected/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-kube-api-access-gh9ks\") pod \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.357361 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-svc\") pod \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\" (UID: \"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5\") " Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.363881 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-kube-api-access-gh9ks" (OuterVolumeSpecName: "kube-api-access-gh9ks") pod "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" (UID: "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5"). InnerVolumeSpecName "kube-api-access-gh9ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.436502 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" (UID: "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.441036 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-config" (OuterVolumeSpecName: "config") pod "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" (UID: "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.442111 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" (UID: "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.442242 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" (UID: "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.443920 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" (UID: "3afa9bc2-f292-4050-b9f2-2c054d8ed0c5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.458622 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.458664 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.458675 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.458683 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.458692 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh9ks\" (UniqueName: \"kubernetes.io/projected/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-kube-api-access-gh9ks\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.458701 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.492898 4799 generic.go:334] "Generic (PLEG): container finished" podID="3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" containerID="3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb" exitCode=0 Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.492986 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" event={"ID":"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5","Type":"ContainerDied","Data":"3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb"} Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.493016 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" event={"ID":"3afa9bc2-f292-4050-b9f2-2c054d8ed0c5","Type":"ContainerDied","Data":"8874e78da064a4f3d93503720862a02cd78f532d28601ce978e5f593d824ae70"} Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.493032 4799 scope.go:117] "RemoveContainer" containerID="3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.493130 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-nwn2f" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.538101 4799 scope.go:117] "RemoveContainer" containerID="acac3cacb1852d0871215875e1702280b7581812d0c21cbcb57e32c22a26256b" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.544160 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.552823 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-nwn2f"] Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.562458 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-nwn2f"] Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.568393 4799 scope.go:117] "RemoveContainer" containerID="3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb" Nov 24 07:06:25 crc kubenswrapper[4799]: E1124 07:06:25.569327 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb\": container with ID starting with 3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb not found: ID does not exist" containerID="3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.569371 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb"} err="failed to get container status \"3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb\": rpc error: code = NotFound desc = could not find container \"3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb\": container with ID starting with 3043f32f3ee02cba8b2f12ab1bce8848ad60c4c405a64c66877dac5cefd1bffb not found: ID does not exist" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.569399 4799 scope.go:117] "RemoveContainer" containerID="acac3cacb1852d0871215875e1702280b7581812d0c21cbcb57e32c22a26256b" Nov 24 07:06:25 crc kubenswrapper[4799]: E1124 07:06:25.569743 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acac3cacb1852d0871215875e1702280b7581812d0c21cbcb57e32c22a26256b\": container with ID starting with acac3cacb1852d0871215875e1702280b7581812d0c21cbcb57e32c22a26256b not found: ID does not exist" containerID="acac3cacb1852d0871215875e1702280b7581812d0c21cbcb57e32c22a26256b" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.569768 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acac3cacb1852d0871215875e1702280b7581812d0c21cbcb57e32c22a26256b"} err="failed to get container status \"acac3cacb1852d0871215875e1702280b7581812d0c21cbcb57e32c22a26256b\": rpc error: code = NotFound desc = could not find container \"acac3cacb1852d0871215875e1702280b7581812d0c21cbcb57e32c22a26256b\": container with ID starting with acac3cacb1852d0871215875e1702280b7581812d0c21cbcb57e32c22a26256b not found: ID does not exist" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.638800 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" path="/var/lib/kubelet/pods/3afa9bc2-f292-4050-b9f2-2c054d8ed0c5/volumes" Nov 24 07:06:25 crc kubenswrapper[4799]: I1124 07:06:25.639435 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" path="/var/lib/kubelet/pods/f633fe14-fd8c-4738-9960-4fff9619c027/volumes" Nov 24 07:06:26 crc kubenswrapper[4799]: I1124 07:06:26.515371 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" containerName="cinder-scheduler" containerID="cri-o://391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f" gracePeriod=30 Nov 24 07:06:26 crc kubenswrapper[4799]: I1124 07:06:26.515445 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" containerName="probe" containerID="cri-o://aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27" gracePeriod=30 Nov 24 07:06:27 crc kubenswrapper[4799]: I1124 07:06:27.533705 4799 generic.go:334] "Generic (PLEG): container finished" podID="cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" containerID="aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27" exitCode=0 Nov 24 07:06:27 crc kubenswrapper[4799]: I1124 07:06:27.534005 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab","Type":"ContainerDied","Data":"aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27"} Nov 24 07:06:28 crc kubenswrapper[4799]: I1124 07:06:28.208349 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:06:28 crc kubenswrapper[4799]: I1124 07:06:28.961432 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:06:28 crc kubenswrapper[4799]: I1124 07:06:28.962166 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:06:29 crc kubenswrapper[4799]: I1124 07:06:29.486204 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.272707 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 07:06:30 crc kubenswrapper[4799]: E1124 07:06:30.273488 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.273506 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api" Nov 24 07:06:30 crc kubenswrapper[4799]: E1124 07:06:30.273533 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api-log" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.273540 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api-log" Nov 24 07:06:30 crc kubenswrapper[4799]: E1124 07:06:30.273553 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" containerName="init" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.273560 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" containerName="init" Nov 24 07:06:30 crc kubenswrapper[4799]: E1124 07:06:30.273598 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" containerName="dnsmasq-dns" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.273605 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" containerName="dnsmasq-dns" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.273830 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3afa9bc2-f292-4050-b9f2-2c054d8ed0c5" containerName="dnsmasq-dns" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.273908 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api-log" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.273922 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f633fe14-fd8c-4738-9960-4fff9619c027" containerName="barbican-api" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.274560 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.276995 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.277318 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-8vq65" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.277697 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.283213 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.368213 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.368295 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf759\" (UniqueName: \"kubernetes.io/projected/76771b43-063e-4bc4-a366-3a13f27e3124-kube-api-access-xf759\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.368322 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config-secret\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.368377 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-combined-ca-bundle\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.397053 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.420153 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.470397 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.470531 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf759\" (UniqueName: \"kubernetes.io/projected/76771b43-063e-4bc4-a366-3a13f27e3124-kube-api-access-xf759\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.470566 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config-secret\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.470672 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-combined-ca-bundle\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.472729 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.492548 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-combined-ca-bundle\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.495670 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf759\" (UniqueName: \"kubernetes.io/projected/76771b43-063e-4bc4-a366-3a13f27e3124-kube-api-access-xf759\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.517367 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config-secret\") pod \"openstackclient\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.526993 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-78cf669d8b-gv6xt"] Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.527307 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-78cf669d8b-gv6xt" podUID="138758e5-99ec-4730-85ec-393117cae96f" containerName="neutron-api" containerID="cri-o://3259793339fb0fed04151354af63bbc8f62939707a77cc329a87cb86d9871b65" gracePeriod=30 Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.527461 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-78cf669d8b-gv6xt" podUID="138758e5-99ec-4730-85ec-393117cae96f" containerName="neutron-httpd" containerID="cri-o://f0d526cab039dc7051e8feec957eed77f1379399fb68a6b9d251d359457ebd40" gracePeriod=30 Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.572837 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data\") pod \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.573138 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrjzj\" (UniqueName: \"kubernetes.io/projected/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-kube-api-access-qrjzj\") pod \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.573279 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-combined-ca-bundle\") pod \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.573364 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-scripts\") pod \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.573404 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data-custom\") pod \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.573452 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-etc-machine-id\") pod \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\" (UID: \"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab\") " Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.574025 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" (UID: "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.579683 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" (UID: "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.584145 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-kube-api-access-qrjzj" (OuterVolumeSpecName: "kube-api-access-qrjzj") pod "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" (UID: "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab"). InnerVolumeSpecName "kube-api-access-qrjzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.591397 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-scripts" (OuterVolumeSpecName: "scripts") pod "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" (UID: "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.598349 4799 generic.go:334] "Generic (PLEG): container finished" podID="cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" containerID="391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f" exitCode=0 Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.598420 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab","Type":"ContainerDied","Data":"391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f"} Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.598460 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd4a9d91-5ccf-41fa-a470-18b4d0c758ab","Type":"ContainerDied","Data":"383c79db7baa1d204c0f2cff05798b094926b21607cfc4c4c9aefa6edca3e7aa"} Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.598485 4799 scope.go:117] "RemoveContainer" containerID="aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.598732 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.603483 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.647372 4799 scope.go:117] "RemoveContainer" containerID="391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.656300 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" (UID: "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.676405 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.676441 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.676455 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.676458 4799 scope.go:117] "RemoveContainer" containerID="aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.676469 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrjzj\" (UniqueName: \"kubernetes.io/projected/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-kube-api-access-qrjzj\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.676586 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:30 crc kubenswrapper[4799]: E1124 07:06:30.686526 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27\": container with ID starting with aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27 not found: ID does not exist" containerID="aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.686580 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27"} err="failed to get container status \"aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27\": rpc error: code = NotFound desc = could not find container \"aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27\": container with ID starting with aa940bb750ecd178107b46f4e42c5073dcd643658911c919519b38426043cc27 not found: ID does not exist" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.686611 4799 scope.go:117] "RemoveContainer" containerID="391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f" Nov 24 07:06:30 crc kubenswrapper[4799]: E1124 07:06:30.688066 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f\": container with ID starting with 391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f not found: ID does not exist" containerID="391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.688132 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f"} err="failed to get container status \"391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f\": rpc error: code = NotFound desc = could not find container \"391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f\": container with ID starting with 391e289273a6f791d9be358810c5addcea5d41af7ae886d94528bb74e7aa6a3f not found: ID does not exist" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.701132 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data" (OuterVolumeSpecName: "config-data") pod "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" (UID: "cd4a9d91-5ccf-41fa-a470-18b4d0c758ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.777937 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.947778 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:06:30 crc kubenswrapper[4799]: I1124 07:06:30.956132 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.013688 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:06:31 crc kubenswrapper[4799]: E1124 07:06:31.014624 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" containerName="cinder-scheduler" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.014648 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" containerName="cinder-scheduler" Nov 24 07:06:31 crc kubenswrapper[4799]: E1124 07:06:31.014702 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" containerName="probe" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.014709 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" containerName="probe" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.015122 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" containerName="cinder-scheduler" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.015176 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" containerName="probe" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.019556 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.023492 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.024366 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.093892 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.094001 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.094045 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb94fabc-dfc3-40cc-9791-0272924692da-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.094063 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgk6w\" (UniqueName: \"kubernetes.io/projected/cb94fabc-dfc3-40cc-9791-0272924692da-kube-api-access-sgk6w\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.094081 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.094111 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.148200 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 07:06:31 crc kubenswrapper[4799]: W1124 07:06:31.157995 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76771b43_063e_4bc4_a366_3a13f27e3124.slice/crio-417c9ffa6e60039e1d67271c6feb7193945f0d4bf91d7b395a731ef96aa63e43 WatchSource:0}: Error finding container 417c9ffa6e60039e1d67271c6feb7193945f0d4bf91d7b395a731ef96aa63e43: Status 404 returned error can't find the container with id 417c9ffa6e60039e1d67271c6feb7193945f0d4bf91d7b395a731ef96aa63e43 Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.195235 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.195576 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb94fabc-dfc3-40cc-9791-0272924692da-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.195596 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgk6w\" (UniqueName: \"kubernetes.io/projected/cb94fabc-dfc3-40cc-9791-0272924692da-kube-api-access-sgk6w\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.195611 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.195641 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.195724 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.198427 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb94fabc-dfc3-40cc-9791-0272924692da-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.201738 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.203063 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.204447 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.206508 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.220452 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgk6w\" (UniqueName: \"kubernetes.io/projected/cb94fabc-dfc3-40cc-9791-0272924692da-kube-api-access-sgk6w\") pod \"cinder-scheduler-0\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.352694 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.617903 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"76771b43-063e-4bc4-a366-3a13f27e3124","Type":"ContainerStarted","Data":"417c9ffa6e60039e1d67271c6feb7193945f0d4bf91d7b395a731ef96aa63e43"} Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.620517 4799 generic.go:334] "Generic (PLEG): container finished" podID="138758e5-99ec-4730-85ec-393117cae96f" containerID="f0d526cab039dc7051e8feec957eed77f1379399fb68a6b9d251d359457ebd40" exitCode=0 Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.620548 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cf669d8b-gv6xt" event={"ID":"138758e5-99ec-4730-85ec-393117cae96f","Type":"ContainerDied","Data":"f0d526cab039dc7051e8feec957eed77f1379399fb68a6b9d251d359457ebd40"} Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.641297 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd4a9d91-5ccf-41fa-a470-18b4d0c758ab" path="/var/lib/kubelet/pods/cd4a9d91-5ccf-41fa-a470-18b4d0c758ab/volumes" Nov 24 07:06:31 crc kubenswrapper[4799]: I1124 07:06:31.867705 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:06:31 crc kubenswrapper[4799]: W1124 07:06:31.882109 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb94fabc_dfc3_40cc_9791_0272924692da.slice/crio-50b3ca3a307bdedf42b2c4ab460d71684336723c36da0f227eb3acde7e00ec22 WatchSource:0}: Error finding container 50b3ca3a307bdedf42b2c4ab460d71684336723c36da0f227eb3acde7e00ec22: Status 404 returned error can't find the container with id 50b3ca3a307bdedf42b2c4ab460d71684336723c36da0f227eb3acde7e00ec22 Nov 24 07:06:32 crc kubenswrapper[4799]: I1124 07:06:32.655906 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb94fabc-dfc3-40cc-9791-0272924692da","Type":"ContainerStarted","Data":"dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc"} Nov 24 07:06:32 crc kubenswrapper[4799]: I1124 07:06:32.656799 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb94fabc-dfc3-40cc-9791-0272924692da","Type":"ContainerStarted","Data":"50b3ca3a307bdedf42b2c4ab460d71684336723c36da0f227eb3acde7e00ec22"} Nov 24 07:06:33 crc kubenswrapper[4799]: I1124 07:06:33.668526 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb94fabc-dfc3-40cc-9791-0272924692da","Type":"ContainerStarted","Data":"af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5"} Nov 24 07:06:33 crc kubenswrapper[4799]: I1124 07:06:33.713506 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.713484007 podStartE2EDuration="3.713484007s" podCreationTimestamp="2025-11-24 07:06:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:33.689048746 +0000 UTC m=+1139.345031220" watchObservedRunningTime="2025-11-24 07:06:33.713484007 +0000 UTC m=+1139.369466481" Nov 24 07:06:34 crc kubenswrapper[4799]: I1124 07:06:34.000502 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.105513 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-9445d5865-fhgvf"] Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.110381 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.114923 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.120821 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.121061 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.131229 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq7j4\" (UniqueName: \"kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-kube-api-access-dq7j4\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.131559 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-combined-ca-bundle\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.131668 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-internal-tls-certs\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.131745 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-log-httpd\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.131820 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-config-data\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.131975 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-run-httpd\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.132247 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-public-tls-certs\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.132321 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-etc-swift\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.137919 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-9445d5865-fhgvf"] Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.189766 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.192269 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="ceilometer-central-agent" containerID="cri-o://77e3d849b1ab7a53dd8b46c090efd860e8595b4933e9efb477e2ce9bac051e9c" gracePeriod=30 Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.193399 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="proxy-httpd" containerID="cri-o://a79bf0ce1dbcde8595fb25f6ca326209543f8a872e833afc9b35f105f1ec00f4" gracePeriod=30 Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.193475 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="sg-core" containerID="cri-o://2e90088664ca828c104ba6d3f4557da6a5523afb97eb2c9138f9a5e9f356b782" gracePeriod=30 Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.193534 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="ceilometer-notification-agent" containerID="cri-o://3572b6c411d077e12bdd5010147fb9798945058a7833556cd2f07866a4bbe975" gracePeriod=30 Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.209041 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.165:3000/\": EOF" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.234725 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-combined-ca-bundle\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.234791 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-internal-tls-certs\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.234822 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-log-httpd\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.234859 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-config-data\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.234902 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-run-httpd\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.234942 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-public-tls-certs\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.234976 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-etc-swift\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.235034 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq7j4\" (UniqueName: \"kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-kube-api-access-dq7j4\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.235683 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-run-httpd\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.235978 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-log-httpd\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.245451 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-public-tls-certs\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.245598 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-combined-ca-bundle\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.245691 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-config-data\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.251082 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-etc-swift\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.256345 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-internal-tls-certs\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.258115 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq7j4\" (UniqueName: \"kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-kube-api-access-dq7j4\") pod \"swift-proxy-9445d5865-fhgvf\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.354048 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.446392 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.711644 4799 generic.go:334] "Generic (PLEG): container finished" podID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerID="a79bf0ce1dbcde8595fb25f6ca326209543f8a872e833afc9b35f105f1ec00f4" exitCode=0 Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.711685 4799 generic.go:334] "Generic (PLEG): container finished" podID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerID="2e90088664ca828c104ba6d3f4557da6a5523afb97eb2c9138f9a5e9f356b782" exitCode=2 Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.711693 4799 generic.go:334] "Generic (PLEG): container finished" podID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerID="77e3d849b1ab7a53dd8b46c090efd860e8595b4933e9efb477e2ce9bac051e9c" exitCode=0 Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.711748 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4c442bc-d661-4294-a156-a0ad37eebc7c","Type":"ContainerDied","Data":"a79bf0ce1dbcde8595fb25f6ca326209543f8a872e833afc9b35f105f1ec00f4"} Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.711785 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4c442bc-d661-4294-a156-a0ad37eebc7c","Type":"ContainerDied","Data":"2e90088664ca828c104ba6d3f4557da6a5523afb97eb2c9138f9a5e9f356b782"} Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.711796 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4c442bc-d661-4294-a156-a0ad37eebc7c","Type":"ContainerDied","Data":"77e3d849b1ab7a53dd8b46c090efd860e8595b4933e9efb477e2ce9bac051e9c"} Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.717301 4799 generic.go:334] "Generic (PLEG): container finished" podID="138758e5-99ec-4730-85ec-393117cae96f" containerID="3259793339fb0fed04151354af63bbc8f62939707a77cc329a87cb86d9871b65" exitCode=0 Nov 24 07:06:36 crc kubenswrapper[4799]: I1124 07:06:36.717367 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cf669d8b-gv6xt" event={"ID":"138758e5-99ec-4730-85ec-393117cae96f","Type":"ContainerDied","Data":"3259793339fb0fed04151354af63bbc8f62939707a77cc329a87cb86d9871b65"} Nov 24 07:06:41 crc kubenswrapper[4799]: I1124 07:06:41.548110 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 07:06:41 crc kubenswrapper[4799]: I1124 07:06:41.803000 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4c442bc-d661-4294-a156-a0ad37eebc7c","Type":"ContainerDied","Data":"3572b6c411d077e12bdd5010147fb9798945058a7833556cd2f07866a4bbe975"} Nov 24 07:06:41 crc kubenswrapper[4799]: I1124 07:06:41.802930 4799 generic.go:334] "Generic (PLEG): container finished" podID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerID="3572b6c411d077e12bdd5010147fb9798945058a7833556cd2f07866a4bbe975" exitCode=0 Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.287020 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-mwx4g"] Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.288739 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mwx4g" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.307900 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mwx4g"] Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.333560 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.366073 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-combined-ca-bundle\") pod \"d4c442bc-d661-4294-a156-a0ad37eebc7c\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.366452 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-config-data\") pod \"d4c442bc-d661-4294-a156-a0ad37eebc7c\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.366563 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-sg-core-conf-yaml\") pod \"d4c442bc-d661-4294-a156-a0ad37eebc7c\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.366660 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw9fx\" (UniqueName: \"kubernetes.io/projected/d4c442bc-d661-4294-a156-a0ad37eebc7c-kube-api-access-fw9fx\") pod \"d4c442bc-d661-4294-a156-a0ad37eebc7c\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.366806 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-run-httpd\") pod \"d4c442bc-d661-4294-a156-a0ad37eebc7c\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.366985 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-log-httpd\") pod \"d4c442bc-d661-4294-a156-a0ad37eebc7c\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.367126 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-scripts\") pod \"d4c442bc-d661-4294-a156-a0ad37eebc7c\" (UID: \"d4c442bc-d661-4294-a156-a0ad37eebc7c\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.370467 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7130f73f-5578-4310-96f3-db076a954990-operator-scripts\") pod \"nova-api-db-create-mwx4g\" (UID: \"7130f73f-5578-4310-96f3-db076a954990\") " pod="openstack/nova-api-db-create-mwx4g" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.370726 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nl9q\" (UniqueName: \"kubernetes.io/projected/7130f73f-5578-4310-96f3-db076a954990-kube-api-access-6nl9q\") pod \"nova-api-db-create-mwx4g\" (UID: \"7130f73f-5578-4310-96f3-db076a954990\") " pod="openstack/nova-api-db-create-mwx4g" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.368230 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d4c442bc-d661-4294-a156-a0ad37eebc7c" (UID: "d4c442bc-d661-4294-a156-a0ad37eebc7c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.372660 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.374259 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d4c442bc-d661-4294-a156-a0ad37eebc7c" (UID: "d4c442bc-d661-4294-a156-a0ad37eebc7c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.383013 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4c442bc-d661-4294-a156-a0ad37eebc7c-kube-api-access-fw9fx" (OuterVolumeSpecName: "kube-api-access-fw9fx") pod "d4c442bc-d661-4294-a156-a0ad37eebc7c" (UID: "d4c442bc-d661-4294-a156-a0ad37eebc7c"). InnerVolumeSpecName "kube-api-access-fw9fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.385545 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-scripts" (OuterVolumeSpecName: "scripts") pod "d4c442bc-d661-4294-a156-a0ad37eebc7c" (UID: "d4c442bc-d661-4294-a156-a0ad37eebc7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.385618 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-z877t"] Nov 24 07:06:42 crc kubenswrapper[4799]: E1124 07:06:42.386120 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="sg-core" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.386137 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="sg-core" Nov 24 07:06:42 crc kubenswrapper[4799]: E1124 07:06:42.386151 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="ceilometer-central-agent" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.386159 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="ceilometer-central-agent" Nov 24 07:06:42 crc kubenswrapper[4799]: E1124 07:06:42.386204 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="ceilometer-notification-agent" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.386227 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="ceilometer-notification-agent" Nov 24 07:06:42 crc kubenswrapper[4799]: E1124 07:06:42.386264 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="proxy-httpd" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.386272 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="proxy-httpd" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.386933 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="ceilometer-notification-agent" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.386970 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="sg-core" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.387000 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="proxy-httpd" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.387009 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" containerName="ceilometer-central-agent" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.388028 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z877t" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.394373 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-z877t"] Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.430412 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d4c442bc-d661-4294-a156-a0ad37eebc7c" (UID: "d4c442bc-d661-4294-a156-a0ad37eebc7c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.435565 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.473802 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nl9q\" (UniqueName: \"kubernetes.io/projected/7130f73f-5578-4310-96f3-db076a954990-kube-api-access-6nl9q\") pod \"nova-api-db-create-mwx4g\" (UID: \"7130f73f-5578-4310-96f3-db076a954990\") " pod="openstack/nova-api-db-create-mwx4g" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.473968 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7130f73f-5578-4310-96f3-db076a954990-operator-scripts\") pod \"nova-api-db-create-mwx4g\" (UID: \"7130f73f-5578-4310-96f3-db076a954990\") " pod="openstack/nova-api-db-create-mwx4g" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.474064 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4c442bc-d661-4294-a156-a0ad37eebc7c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.474074 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.474093 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.474104 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw9fx\" (UniqueName: \"kubernetes.io/projected/d4c442bc-d661-4294-a156-a0ad37eebc7c-kube-api-access-fw9fx\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.474794 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7130f73f-5578-4310-96f3-db076a954990-operator-scripts\") pod \"nova-api-db-create-mwx4g\" (UID: \"7130f73f-5578-4310-96f3-db076a954990\") " pod="openstack/nova-api-db-create-mwx4g" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.495693 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4c442bc-d661-4294-a156-a0ad37eebc7c" (UID: "d4c442bc-d661-4294-a156-a0ad37eebc7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.501108 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nl9q\" (UniqueName: \"kubernetes.io/projected/7130f73f-5578-4310-96f3-db076a954990-kube-api-access-6nl9q\") pod \"nova-api-db-create-mwx4g\" (UID: \"7130f73f-5578-4310-96f3-db076a954990\") " pod="openstack/nova-api-db-create-mwx4g" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.518875 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-722e-account-create-5cz9z"] Nov 24 07:06:42 crc kubenswrapper[4799]: E1124 07:06:42.520104 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="138758e5-99ec-4730-85ec-393117cae96f" containerName="neutron-api" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.520189 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="138758e5-99ec-4730-85ec-393117cae96f" containerName="neutron-api" Nov 24 07:06:42 crc kubenswrapper[4799]: E1124 07:06:42.520279 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="138758e5-99ec-4730-85ec-393117cae96f" containerName="neutron-httpd" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.520337 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="138758e5-99ec-4730-85ec-393117cae96f" containerName="neutron-httpd" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.520607 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="138758e5-99ec-4730-85ec-393117cae96f" containerName="neutron-httpd" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.520723 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="138758e5-99ec-4730-85ec-393117cae96f" containerName="neutron-api" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.521747 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-722e-account-create-5cz9z" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.525149 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-config-data" (OuterVolumeSpecName: "config-data") pod "d4c442bc-d661-4294-a156-a0ad37eebc7c" (UID: "d4c442bc-d661-4294-a156-a0ad37eebc7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.525402 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.534195 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-722e-account-create-5cz9z"] Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.575413 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-config\") pod \"138758e5-99ec-4730-85ec-393117cae96f\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.576471 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-httpd-config\") pod \"138758e5-99ec-4730-85ec-393117cae96f\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.576555 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-ovndb-tls-certs\") pod \"138758e5-99ec-4730-85ec-393117cae96f\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.576708 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-combined-ca-bundle\") pod \"138758e5-99ec-4730-85ec-393117cae96f\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.576826 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6kqr\" (UniqueName: \"kubernetes.io/projected/138758e5-99ec-4730-85ec-393117cae96f-kube-api-access-f6kqr\") pod \"138758e5-99ec-4730-85ec-393117cae96f\" (UID: \"138758e5-99ec-4730-85ec-393117cae96f\") " Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.577185 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-operator-scripts\") pod \"nova-cell0-db-create-z877t\" (UID: \"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed\") " pod="openstack/nova-cell0-db-create-z877t" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.577306 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5mb9\" (UniqueName: \"kubernetes.io/projected/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-kube-api-access-m5mb9\") pod \"nova-cell0-db-create-z877t\" (UID: \"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed\") " pod="openstack/nova-cell0-db-create-z877t" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.577437 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.577505 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c442bc-d661-4294-a156-a0ad37eebc7c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.580902 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-mkgcm"] Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.581119 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "138758e5-99ec-4730-85ec-393117cae96f" (UID: "138758e5-99ec-4730-85ec-393117cae96f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.583572 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-mkgcm" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.601118 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/138758e5-99ec-4730-85ec-393117cae96f-kube-api-access-f6kqr" (OuterVolumeSpecName: "kube-api-access-f6kqr") pod "138758e5-99ec-4730-85ec-393117cae96f" (UID: "138758e5-99ec-4730-85ec-393117cae96f"). InnerVolumeSpecName "kube-api-access-f6kqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.601668 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-mkgcm"] Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.632023 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mwx4g" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.640400 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-config" (OuterVolumeSpecName: "config") pod "138758e5-99ec-4730-85ec-393117cae96f" (UID: "138758e5-99ec-4730-85ec-393117cae96f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.643978 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-9445d5865-fhgvf"] Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.655261 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "138758e5-99ec-4730-85ec-393117cae96f" (UID: "138758e5-99ec-4730-85ec-393117cae96f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.681493 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5mb9\" (UniqueName: \"kubernetes.io/projected/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-kube-api-access-m5mb9\") pod \"nova-cell0-db-create-z877t\" (UID: \"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed\") " pod="openstack/nova-cell0-db-create-z877t" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.681885 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9d5k\" (UniqueName: \"kubernetes.io/projected/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-kube-api-access-k9d5k\") pod \"nova-api-722e-account-create-5cz9z\" (UID: \"3dd8bba3-489f-442a-97a4-47d2ebffdc9b\") " pod="openstack/nova-api-722e-account-create-5cz9z" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.682012 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l78ln\" (UniqueName: \"kubernetes.io/projected/e733dc01-8d9d-4a0b-b2e7-e18147562e11-kube-api-access-l78ln\") pod \"nova-cell1-db-create-mkgcm\" (UID: \"e733dc01-8d9d-4a0b-b2e7-e18147562e11\") " pod="openstack/nova-cell1-db-create-mkgcm" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.682113 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-operator-scripts\") pod \"nova-api-722e-account-create-5cz9z\" (UID: \"3dd8bba3-489f-442a-97a4-47d2ebffdc9b\") " pod="openstack/nova-api-722e-account-create-5cz9z" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.682252 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-operator-scripts\") pod \"nova-cell0-db-create-z877t\" (UID: \"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed\") " pod="openstack/nova-cell0-db-create-z877t" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.682362 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e733dc01-8d9d-4a0b-b2e7-e18147562e11-operator-scripts\") pod \"nova-cell1-db-create-mkgcm\" (UID: \"e733dc01-8d9d-4a0b-b2e7-e18147562e11\") " pod="openstack/nova-cell1-db-create-mkgcm" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.682568 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.682631 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6kqr\" (UniqueName: \"kubernetes.io/projected/138758e5-99ec-4730-85ec-393117cae96f-kube-api-access-f6kqr\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.682769 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.682827 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.683999 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-operator-scripts\") pod \"nova-cell0-db-create-z877t\" (UID: \"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed\") " pod="openstack/nova-cell0-db-create-z877t" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.694761 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-00c8-account-create-w4gf7"] Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.697127 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-00c8-account-create-w4gf7" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.699930 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.707352 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-00c8-account-create-w4gf7"] Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.708576 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5mb9\" (UniqueName: \"kubernetes.io/projected/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-kube-api-access-m5mb9\") pod \"nova-cell0-db-create-z877t\" (UID: \"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed\") " pod="openstack/nova-cell0-db-create-z877t" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.716255 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z877t" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.736680 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "138758e5-99ec-4730-85ec-393117cae96f" (UID: "138758e5-99ec-4730-85ec-393117cae96f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.786905 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsw6z\" (UniqueName: \"kubernetes.io/projected/04ddaef5-8849-4fa9-a0b9-f5e009dff420-kube-api-access-dsw6z\") pod \"nova-cell0-00c8-account-create-w4gf7\" (UID: \"04ddaef5-8849-4fa9-a0b9-f5e009dff420\") " pod="openstack/nova-cell0-00c8-account-create-w4gf7" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.787255 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9d5k\" (UniqueName: \"kubernetes.io/projected/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-kube-api-access-k9d5k\") pod \"nova-api-722e-account-create-5cz9z\" (UID: \"3dd8bba3-489f-442a-97a4-47d2ebffdc9b\") " pod="openstack/nova-api-722e-account-create-5cz9z" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.787298 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l78ln\" (UniqueName: \"kubernetes.io/projected/e733dc01-8d9d-4a0b-b2e7-e18147562e11-kube-api-access-l78ln\") pod \"nova-cell1-db-create-mkgcm\" (UID: \"e733dc01-8d9d-4a0b-b2e7-e18147562e11\") " pod="openstack/nova-cell1-db-create-mkgcm" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.787318 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-operator-scripts\") pod \"nova-api-722e-account-create-5cz9z\" (UID: \"3dd8bba3-489f-442a-97a4-47d2ebffdc9b\") " pod="openstack/nova-api-722e-account-create-5cz9z" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.787354 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04ddaef5-8849-4fa9-a0b9-f5e009dff420-operator-scripts\") pod \"nova-cell0-00c8-account-create-w4gf7\" (UID: \"04ddaef5-8849-4fa9-a0b9-f5e009dff420\") " pod="openstack/nova-cell0-00c8-account-create-w4gf7" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.787391 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e733dc01-8d9d-4a0b-b2e7-e18147562e11-operator-scripts\") pod \"nova-cell1-db-create-mkgcm\" (UID: \"e733dc01-8d9d-4a0b-b2e7-e18147562e11\") " pod="openstack/nova-cell1-db-create-mkgcm" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.787453 4799 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/138758e5-99ec-4730-85ec-393117cae96f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.788209 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e733dc01-8d9d-4a0b-b2e7-e18147562e11-operator-scripts\") pod \"nova-cell1-db-create-mkgcm\" (UID: \"e733dc01-8d9d-4a0b-b2e7-e18147562e11\") " pod="openstack/nova-cell1-db-create-mkgcm" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.788270 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-operator-scripts\") pod \"nova-api-722e-account-create-5cz9z\" (UID: \"3dd8bba3-489f-442a-97a4-47d2ebffdc9b\") " pod="openstack/nova-api-722e-account-create-5cz9z" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.807105 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l78ln\" (UniqueName: \"kubernetes.io/projected/e733dc01-8d9d-4a0b-b2e7-e18147562e11-kube-api-access-l78ln\") pod \"nova-cell1-db-create-mkgcm\" (UID: \"e733dc01-8d9d-4a0b-b2e7-e18147562e11\") " pod="openstack/nova-cell1-db-create-mkgcm" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.807286 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9d5k\" (UniqueName: \"kubernetes.io/projected/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-kube-api-access-k9d5k\") pod \"nova-api-722e-account-create-5cz9z\" (UID: \"3dd8bba3-489f-442a-97a4-47d2ebffdc9b\") " pod="openstack/nova-api-722e-account-create-5cz9z" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.836808 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-9445d5865-fhgvf" event={"ID":"b639ad0e-8438-4a36-a776-ee53b6f0c080","Type":"ContainerStarted","Data":"b83ebe9c11737bdf41487c483cafea50aac2f9991b0f568474b028a6b4d951a4"} Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.841589 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4c442bc-d661-4294-a156-a0ad37eebc7c","Type":"ContainerDied","Data":"c90f92146c1825de25aa9c53ca4f215be6b89d17999a5a712d9db4048c58eed1"} Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.841643 4799 scope.go:117] "RemoveContainer" containerID="a79bf0ce1dbcde8595fb25f6ca326209543f8a872e833afc9b35f105f1ec00f4" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.841802 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.845128 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-722e-account-create-5cz9z" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.856474 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"76771b43-063e-4bc4-a366-3a13f27e3124","Type":"ContainerStarted","Data":"1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259"} Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.886117 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78cf669d8b-gv6xt" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.889727 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsw6z\" (UniqueName: \"kubernetes.io/projected/04ddaef5-8849-4fa9-a0b9-f5e009dff420-kube-api-access-dsw6z\") pod \"nova-cell0-00c8-account-create-w4gf7\" (UID: \"04ddaef5-8849-4fa9-a0b9-f5e009dff420\") " pod="openstack/nova-cell0-00c8-account-create-w4gf7" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.890120 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04ddaef5-8849-4fa9-a0b9-f5e009dff420-operator-scripts\") pod \"nova-cell0-00c8-account-create-w4gf7\" (UID: \"04ddaef5-8849-4fa9-a0b9-f5e009dff420\") " pod="openstack/nova-cell0-00c8-account-create-w4gf7" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.893719 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04ddaef5-8849-4fa9-a0b9-f5e009dff420-operator-scripts\") pod \"nova-cell0-00c8-account-create-w4gf7\" (UID: \"04ddaef5-8849-4fa9-a0b9-f5e009dff420\") " pod="openstack/nova-cell0-00c8-account-create-w4gf7" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.910703 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-mkgcm" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.911202 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.237051984 podStartE2EDuration="12.91117774s" podCreationTimestamp="2025-11-24 07:06:30 +0000 UTC" firstStartedPulling="2025-11-24 07:06:31.160145754 +0000 UTC m=+1136.816128228" lastFinishedPulling="2025-11-24 07:06:41.83427151 +0000 UTC m=+1147.490253984" observedRunningTime="2025-11-24 07:06:42.893042987 +0000 UTC m=+1148.549025461" watchObservedRunningTime="2025-11-24 07:06:42.91117774 +0000 UTC m=+1148.567160234" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.886028 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78cf669d8b-gv6xt" event={"ID":"138758e5-99ec-4730-85ec-393117cae96f","Type":"ContainerDied","Data":"1e25741d3e9fc033beede695361855121a2308d99f8a2c0df4b1badd97b27c2d"} Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.918964 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-25d6-account-create-r6lpd"] Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.920156 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-25d6-account-create-r6lpd"] Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.920319 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-25d6-account-create-r6lpd" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.924394 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.924861 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsw6z\" (UniqueName: \"kubernetes.io/projected/04ddaef5-8849-4fa9-a0b9-f5e009dff420-kube-api-access-dsw6z\") pod \"nova-cell0-00c8-account-create-w4gf7\" (UID: \"04ddaef5-8849-4fa9-a0b9-f5e009dff420\") " pod="openstack/nova-cell0-00c8-account-create-w4gf7" Nov 24 07:06:42 crc kubenswrapper[4799]: I1124 07:06:42.949102 4799 scope.go:117] "RemoveContainer" containerID="2e90088664ca828c104ba6d3f4557da6a5523afb97eb2c9138f9a5e9f356b782" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.084756 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-00c8-account-create-w4gf7" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.088108 4799 scope.go:117] "RemoveContainer" containerID="3572b6c411d077e12bdd5010147fb9798945058a7833556cd2f07866a4bbe975" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.097167 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpfw6\" (UniqueName: \"kubernetes.io/projected/5cde9df8-eba6-48ec-bc68-ed1fae49df81-kube-api-access-xpfw6\") pod \"nova-cell1-25d6-account-create-r6lpd\" (UID: \"5cde9df8-eba6-48ec-bc68-ed1fae49df81\") " pod="openstack/nova-cell1-25d6-account-create-r6lpd" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.097226 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cde9df8-eba6-48ec-bc68-ed1fae49df81-operator-scripts\") pod \"nova-cell1-25d6-account-create-r6lpd\" (UID: \"5cde9df8-eba6-48ec-bc68-ed1fae49df81\") " pod="openstack/nova-cell1-25d6-account-create-r6lpd" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.097430 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-78cf669d8b-gv6xt"] Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.111107 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-78cf669d8b-gv6xt"] Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.160603 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.208081 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpfw6\" (UniqueName: \"kubernetes.io/projected/5cde9df8-eba6-48ec-bc68-ed1fae49df81-kube-api-access-xpfw6\") pod \"nova-cell1-25d6-account-create-r6lpd\" (UID: \"5cde9df8-eba6-48ec-bc68-ed1fae49df81\") " pod="openstack/nova-cell1-25d6-account-create-r6lpd" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.211785 4799 scope.go:117] "RemoveContainer" containerID="77e3d849b1ab7a53dd8b46c090efd860e8595b4933e9efb477e2ce9bac051e9c" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.213896 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cde9df8-eba6-48ec-bc68-ed1fae49df81-operator-scripts\") pod \"nova-cell1-25d6-account-create-r6lpd\" (UID: \"5cde9df8-eba6-48ec-bc68-ed1fae49df81\") " pod="openstack/nova-cell1-25d6-account-create-r6lpd" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.214109 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.214782 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cde9df8-eba6-48ec-bc68-ed1fae49df81-operator-scripts\") pod \"nova-cell1-25d6-account-create-r6lpd\" (UID: \"5cde9df8-eba6-48ec-bc68-ed1fae49df81\") " pod="openstack/nova-cell1-25d6-account-create-r6lpd" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.231416 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.232438 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpfw6\" (UniqueName: \"kubernetes.io/projected/5cde9df8-eba6-48ec-bc68-ed1fae49df81-kube-api-access-xpfw6\") pod \"nova-cell1-25d6-account-create-r6lpd\" (UID: \"5cde9df8-eba6-48ec-bc68-ed1fae49df81\") " pod="openstack/nova-cell1-25d6-account-create-r6lpd" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.233679 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.238505 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.245443 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.247433 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:06:43 crc kubenswrapper[4799]: W1124 07:06:43.265826 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7130f73f_5578_4310_96f3_db076a954990.slice/crio-3fefb02d2aefdc2fe255b95879d12e6407cca183fa8c10bfefb403035e3569e3 WatchSource:0}: Error finding container 3fefb02d2aefdc2fe255b95879d12e6407cca183fa8c10bfefb403035e3569e3: Status 404 returned error can't find the container with id 3fefb02d2aefdc2fe255b95879d12e6407cca183fa8c10bfefb403035e3569e3 Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.267607 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mwx4g"] Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.288263 4799 scope.go:117] "RemoveContainer" containerID="f0d526cab039dc7051e8feec957eed77f1379399fb68a6b9d251d359457ebd40" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.341194 4799 scope.go:117] "RemoveContainer" containerID="3259793339fb0fed04151354af63bbc8f62939707a77cc329a87cb86d9871b65" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.419683 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-config-data\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.419730 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-log-httpd\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.419752 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.419820 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-scripts\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.419861 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-run-httpd\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.419903 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jcmj\" (UniqueName: \"kubernetes.io/projected/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-kube-api-access-8jcmj\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.419920 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.459408 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-z877t"] Nov 24 07:06:43 crc kubenswrapper[4799]: W1124 07:06:43.464608 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod558a22c8_97cd_48ab_b7f5_fdd5ea96a5ed.slice/crio-476830d22857f9fe260687edb887cfa22bb84f9b6dcbd2b10787126cad9ffa8c WatchSource:0}: Error finding container 476830d22857f9fe260687edb887cfa22bb84f9b6dcbd2b10787126cad9ffa8c: Status 404 returned error can't find the container with id 476830d22857f9fe260687edb887cfa22bb84f9b6dcbd2b10787126cad9ffa8c Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.476633 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-25d6-account-create-r6lpd" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.521308 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-scripts\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.521349 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-run-httpd\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.521416 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jcmj\" (UniqueName: \"kubernetes.io/projected/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-kube-api-access-8jcmj\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.521434 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.521492 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-config-data\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.521532 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-log-httpd\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.521552 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.528287 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-run-httpd\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.528725 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-log-httpd\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.529532 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.533452 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.537977 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-scripts\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.540840 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-config-data\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.554306 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jcmj\" (UniqueName: \"kubernetes.io/projected/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-kube-api-access-8jcmj\") pod \"ceilometer-0\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.597602 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.620098 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-mkgcm"] Nov 24 07:06:43 crc kubenswrapper[4799]: W1124 07:06:43.625743 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode733dc01_8d9d_4a0b_b2e7_e18147562e11.slice/crio-ca3a4995cac84e1f8ca5ef118d5b97f96e61afc52a31a911b4f76f0881c200c8 WatchSource:0}: Error finding container ca3a4995cac84e1f8ca5ef118d5b97f96e61afc52a31a911b4f76f0881c200c8: Status 404 returned error can't find the container with id ca3a4995cac84e1f8ca5ef118d5b97f96e61afc52a31a911b4f76f0881c200c8 Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.641549 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="138758e5-99ec-4730-85ec-393117cae96f" path="/var/lib/kubelet/pods/138758e5-99ec-4730-85ec-393117cae96f/volumes" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.642201 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4c442bc-d661-4294-a156-a0ad37eebc7c" path="/var/lib/kubelet/pods/d4c442bc-d661-4294-a156-a0ad37eebc7c/volumes" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.642988 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-722e-account-create-5cz9z"] Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.782662 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-00c8-account-create-w4gf7"] Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.921201 4799 generic.go:334] "Generic (PLEG): container finished" podID="7130f73f-5578-4310-96f3-db076a954990" containerID="1bae0067729d54c8ab8a099e80fb824001f7aaa34386e63f9c99d92e5d77407c" exitCode=0 Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.922019 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mwx4g" event={"ID":"7130f73f-5578-4310-96f3-db076a954990","Type":"ContainerDied","Data":"1bae0067729d54c8ab8a099e80fb824001f7aaa34386e63f9c99d92e5d77407c"} Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.922046 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mwx4g" event={"ID":"7130f73f-5578-4310-96f3-db076a954990","Type":"ContainerStarted","Data":"3fefb02d2aefdc2fe255b95879d12e6407cca183fa8c10bfefb403035e3569e3"} Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.925746 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z877t" event={"ID":"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed","Type":"ContainerStarted","Data":"525cb566c1a708ef724e8e60124f3abc876a55149f6adbf23fede6052a7b4b2f"} Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.925791 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z877t" event={"ID":"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed","Type":"ContainerStarted","Data":"476830d22857f9fe260687edb887cfa22bb84f9b6dcbd2b10787126cad9ffa8c"} Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.928092 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-mkgcm" event={"ID":"e733dc01-8d9d-4a0b-b2e7-e18147562e11","Type":"ContainerStarted","Data":"ca3a4995cac84e1f8ca5ef118d5b97f96e61afc52a31a911b4f76f0881c200c8"} Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.933052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-722e-account-create-5cz9z" event={"ID":"3dd8bba3-489f-442a-97a4-47d2ebffdc9b","Type":"ContainerStarted","Data":"c483e2cb2fdc62e52cbf180ed20c8ea4ec9f9d4de0c4cc2e42f7922b74507136"} Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.953762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-00c8-account-create-w4gf7" event={"ID":"04ddaef5-8849-4fa9-a0b9-f5e009dff420","Type":"ContainerStarted","Data":"62cafdd129bdd495bc91cf610c9b7de9678a65a3adcfc244975b53cf904d5026"} Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.960374 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-z877t" podStartSLOduration=1.960359687 podStartE2EDuration="1.960359687s" podCreationTimestamp="2025-11-24 07:06:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:43.950758906 +0000 UTC m=+1149.606741400" watchObservedRunningTime="2025-11-24 07:06:43.960359687 +0000 UTC m=+1149.616342161" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.981043 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-9445d5865-fhgvf" event={"ID":"b639ad0e-8438-4a36-a776-ee53b6f0c080","Type":"ContainerStarted","Data":"f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b"} Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.981473 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-9445d5865-fhgvf" event={"ID":"b639ad0e-8438-4a36-a776-ee53b6f0c080","Type":"ContainerStarted","Data":"2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a"} Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.981537 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:43 crc kubenswrapper[4799]: I1124 07:06:43.981554 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.008738 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-9445d5865-fhgvf" podStartSLOduration=8.008720305 podStartE2EDuration="8.008720305s" podCreationTimestamp="2025-11-24 07:06:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:06:44.006681488 +0000 UTC m=+1149.662663962" watchObservedRunningTime="2025-11-24 07:06:44.008720305 +0000 UTC m=+1149.664702779" Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.111393 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-25d6-account-create-r6lpd"] Nov 24 07:06:44 crc kubenswrapper[4799]: W1124 07:06:44.113762 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cde9df8_eba6_48ec_bc68_ed1fae49df81.slice/crio-27fca7539965822ffb52d2fd4c09236b719b358427ec02abf084153067edd29a WatchSource:0}: Error finding container 27fca7539965822ffb52d2fd4c09236b719b358427ec02abf084153067edd29a: Status 404 returned error can't find the container with id 27fca7539965822ffb52d2fd4c09236b719b358427ec02abf084153067edd29a Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.247802 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:44 crc kubenswrapper[4799]: W1124 07:06:44.254450 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8c58e1a_1d91_4f16_843a_bd6f81a03d7b.slice/crio-cf4f4a234f688bdd7d36b9245197a806bfdda60c4553c7fd0012c9cc8cb32922 WatchSource:0}: Error finding container cf4f4a234f688bdd7d36b9245197a806bfdda60c4553c7fd0012c9cc8cb32922: Status 404 returned error can't find the container with id cf4f4a234f688bdd7d36b9245197a806bfdda60c4553c7fd0012c9cc8cb32922 Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.910740 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.991113 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b","Type":"ContainerStarted","Data":"cf4f4a234f688bdd7d36b9245197a806bfdda60c4553c7fd0012c9cc8cb32922"} Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.994342 4799 generic.go:334] "Generic (PLEG): container finished" podID="5cde9df8-eba6-48ec-bc68-ed1fae49df81" containerID="bf402ab86034c98ced891cf3177f170ae03981592178d303ce73115272760cb3" exitCode=0 Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.994553 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-25d6-account-create-r6lpd" event={"ID":"5cde9df8-eba6-48ec-bc68-ed1fae49df81","Type":"ContainerDied","Data":"bf402ab86034c98ced891cf3177f170ae03981592178d303ce73115272760cb3"} Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.994613 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-25d6-account-create-r6lpd" event={"ID":"5cde9df8-eba6-48ec-bc68-ed1fae49df81","Type":"ContainerStarted","Data":"27fca7539965822ffb52d2fd4c09236b719b358427ec02abf084153067edd29a"} Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.995803 4799 generic.go:334] "Generic (PLEG): container finished" podID="04ddaef5-8849-4fa9-a0b9-f5e009dff420" containerID="b1cbe7bfd60ea9dda2a4263ac59bad2dfc42f541a2a9046fdc4fb6a455efc23e" exitCode=0 Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.996081 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-00c8-account-create-w4gf7" event={"ID":"04ddaef5-8849-4fa9-a0b9-f5e009dff420","Type":"ContainerDied","Data":"b1cbe7bfd60ea9dda2a4263ac59bad2dfc42f541a2a9046fdc4fb6a455efc23e"} Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.998176 4799 generic.go:334] "Generic (PLEG): container finished" podID="558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed" containerID="525cb566c1a708ef724e8e60124f3abc876a55149f6adbf23fede6052a7b4b2f" exitCode=0 Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.998324 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z877t" event={"ID":"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed","Type":"ContainerDied","Data":"525cb566c1a708ef724e8e60124f3abc876a55149f6adbf23fede6052a7b4b2f"} Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.999583 4799 generic.go:334] "Generic (PLEG): container finished" podID="e733dc01-8d9d-4a0b-b2e7-e18147562e11" containerID="482794ed63a9b329b5550fafc91a694f09f866c3013f86deef50df49ef6c173e" exitCode=0 Nov 24 07:06:44 crc kubenswrapper[4799]: I1124 07:06:44.999738 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-mkgcm" event={"ID":"e733dc01-8d9d-4a0b-b2e7-e18147562e11","Type":"ContainerDied","Data":"482794ed63a9b329b5550fafc91a694f09f866c3013f86deef50df49ef6c173e"} Nov 24 07:06:45 crc kubenswrapper[4799]: I1124 07:06:45.000980 4799 generic.go:334] "Generic (PLEG): container finished" podID="3dd8bba3-489f-442a-97a4-47d2ebffdc9b" containerID="caf4f5eb8dd9a9168446c28bcc88c19555451498f7a234dc9f24098b6462bfd0" exitCode=0 Nov 24 07:06:45 crc kubenswrapper[4799]: I1124 07:06:45.001955 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-722e-account-create-5cz9z" event={"ID":"3dd8bba3-489f-442a-97a4-47d2ebffdc9b","Type":"ContainerDied","Data":"caf4f5eb8dd9a9168446c28bcc88c19555451498f7a234dc9f24098b6462bfd0"} Nov 24 07:06:45 crc kubenswrapper[4799]: I1124 07:06:45.450088 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mwx4g" Nov 24 07:06:45 crc kubenswrapper[4799]: I1124 07:06:45.596449 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nl9q\" (UniqueName: \"kubernetes.io/projected/7130f73f-5578-4310-96f3-db076a954990-kube-api-access-6nl9q\") pod \"7130f73f-5578-4310-96f3-db076a954990\" (UID: \"7130f73f-5578-4310-96f3-db076a954990\") " Nov 24 07:06:45 crc kubenswrapper[4799]: I1124 07:06:45.596669 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7130f73f-5578-4310-96f3-db076a954990-operator-scripts\") pod \"7130f73f-5578-4310-96f3-db076a954990\" (UID: \"7130f73f-5578-4310-96f3-db076a954990\") " Nov 24 07:06:45 crc kubenswrapper[4799]: I1124 07:06:45.597791 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7130f73f-5578-4310-96f3-db076a954990-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7130f73f-5578-4310-96f3-db076a954990" (UID: "7130f73f-5578-4310-96f3-db076a954990"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:45 crc kubenswrapper[4799]: I1124 07:06:45.603704 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7130f73f-5578-4310-96f3-db076a954990-kube-api-access-6nl9q" (OuterVolumeSpecName: "kube-api-access-6nl9q") pod "7130f73f-5578-4310-96f3-db076a954990" (UID: "7130f73f-5578-4310-96f3-db076a954990"). InnerVolumeSpecName "kube-api-access-6nl9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:45 crc kubenswrapper[4799]: I1124 07:06:45.698718 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nl9q\" (UniqueName: \"kubernetes.io/projected/7130f73f-5578-4310-96f3-db076a954990-kube-api-access-6nl9q\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:45 crc kubenswrapper[4799]: I1124 07:06:45.698747 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7130f73f-5578-4310-96f3-db076a954990-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.017053 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b","Type":"ContainerStarted","Data":"b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68"} Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.018975 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mwx4g" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.018968 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mwx4g" event={"ID":"7130f73f-5578-4310-96f3-db076a954990","Type":"ContainerDied","Data":"3fefb02d2aefdc2fe255b95879d12e6407cca183fa8c10bfefb403035e3569e3"} Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.019046 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fefb02d2aefdc2fe255b95879d12e6407cca183fa8c10bfefb403035e3569e3" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.457382 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-00c8-account-create-w4gf7" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.607001 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z877t" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.625434 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsw6z\" (UniqueName: \"kubernetes.io/projected/04ddaef5-8849-4fa9-a0b9-f5e009dff420-kube-api-access-dsw6z\") pod \"04ddaef5-8849-4fa9-a0b9-f5e009dff420\" (UID: \"04ddaef5-8849-4fa9-a0b9-f5e009dff420\") " Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.625501 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5mb9\" (UniqueName: \"kubernetes.io/projected/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-kube-api-access-m5mb9\") pod \"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed\" (UID: \"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed\") " Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.625539 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-operator-scripts\") pod \"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed\" (UID: \"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed\") " Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.625760 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04ddaef5-8849-4fa9-a0b9-f5e009dff420-operator-scripts\") pod \"04ddaef5-8849-4fa9-a0b9-f5e009dff420\" (UID: \"04ddaef5-8849-4fa9-a0b9-f5e009dff420\") " Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.627159 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04ddaef5-8849-4fa9-a0b9-f5e009dff420-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "04ddaef5-8849-4fa9-a0b9-f5e009dff420" (UID: "04ddaef5-8849-4fa9-a0b9-f5e009dff420"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.627544 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed" (UID: "558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.632480 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04ddaef5-8849-4fa9-a0b9-f5e009dff420-kube-api-access-dsw6z" (OuterVolumeSpecName: "kube-api-access-dsw6z") pod "04ddaef5-8849-4fa9-a0b9-f5e009dff420" (UID: "04ddaef5-8849-4fa9-a0b9-f5e009dff420"). InnerVolumeSpecName "kube-api-access-dsw6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.636338 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-kube-api-access-m5mb9" (OuterVolumeSpecName: "kube-api-access-m5mb9") pod "558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed" (UID: "558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed"). InnerVolumeSpecName "kube-api-access-m5mb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.639864 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-25d6-account-create-r6lpd" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.663562 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-mkgcm" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.725053 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-722e-account-create-5cz9z" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.727406 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9d5k\" (UniqueName: \"kubernetes.io/projected/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-kube-api-access-k9d5k\") pod \"3dd8bba3-489f-442a-97a4-47d2ebffdc9b\" (UID: \"3dd8bba3-489f-442a-97a4-47d2ebffdc9b\") " Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.727540 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-operator-scripts\") pod \"3dd8bba3-489f-442a-97a4-47d2ebffdc9b\" (UID: \"3dd8bba3-489f-442a-97a4-47d2ebffdc9b\") " Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.727582 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cde9df8-eba6-48ec-bc68-ed1fae49df81-operator-scripts\") pod \"5cde9df8-eba6-48ec-bc68-ed1fae49df81\" (UID: \"5cde9df8-eba6-48ec-bc68-ed1fae49df81\") " Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.727643 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l78ln\" (UniqueName: \"kubernetes.io/projected/e733dc01-8d9d-4a0b-b2e7-e18147562e11-kube-api-access-l78ln\") pod \"e733dc01-8d9d-4a0b-b2e7-e18147562e11\" (UID: \"e733dc01-8d9d-4a0b-b2e7-e18147562e11\") " Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.727673 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e733dc01-8d9d-4a0b-b2e7-e18147562e11-operator-scripts\") pod \"e733dc01-8d9d-4a0b-b2e7-e18147562e11\" (UID: \"e733dc01-8d9d-4a0b-b2e7-e18147562e11\") " Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.727711 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpfw6\" (UniqueName: \"kubernetes.io/projected/5cde9df8-eba6-48ec-bc68-ed1fae49df81-kube-api-access-xpfw6\") pod \"5cde9df8-eba6-48ec-bc68-ed1fae49df81\" (UID: \"5cde9df8-eba6-48ec-bc68-ed1fae49df81\") " Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.728131 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3dd8bba3-489f-442a-97a4-47d2ebffdc9b" (UID: "3dd8bba3-489f-442a-97a4-47d2ebffdc9b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.728650 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cde9df8-eba6-48ec-bc68-ed1fae49df81-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5cde9df8-eba6-48ec-bc68-ed1fae49df81" (UID: "5cde9df8-eba6-48ec-bc68-ed1fae49df81"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.728660 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e733dc01-8d9d-4a0b-b2e7-e18147562e11-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e733dc01-8d9d-4a0b-b2e7-e18147562e11" (UID: "e733dc01-8d9d-4a0b-b2e7-e18147562e11"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.730376 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsw6z\" (UniqueName: \"kubernetes.io/projected/04ddaef5-8849-4fa9-a0b9-f5e009dff420-kube-api-access-dsw6z\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.730400 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5mb9\" (UniqueName: \"kubernetes.io/projected/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-kube-api-access-m5mb9\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.730410 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cde9df8-eba6-48ec-bc68-ed1fae49df81-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.730419 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.730428 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e733dc01-8d9d-4a0b-b2e7-e18147562e11-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.730436 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04ddaef5-8849-4fa9-a0b9-f5e009dff420-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.732294 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cde9df8-eba6-48ec-bc68-ed1fae49df81-kube-api-access-xpfw6" (OuterVolumeSpecName: "kube-api-access-xpfw6") pod "5cde9df8-eba6-48ec-bc68-ed1fae49df81" (UID: "5cde9df8-eba6-48ec-bc68-ed1fae49df81"). InnerVolumeSpecName "kube-api-access-xpfw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.735049 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-kube-api-access-k9d5k" (OuterVolumeSpecName: "kube-api-access-k9d5k") pod "3dd8bba3-489f-442a-97a4-47d2ebffdc9b" (UID: "3dd8bba3-489f-442a-97a4-47d2ebffdc9b"). InnerVolumeSpecName "kube-api-access-k9d5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.735159 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e733dc01-8d9d-4a0b-b2e7-e18147562e11-kube-api-access-l78ln" (OuterVolumeSpecName: "kube-api-access-l78ln") pod "e733dc01-8d9d-4a0b-b2e7-e18147562e11" (UID: "e733dc01-8d9d-4a0b-b2e7-e18147562e11"). InnerVolumeSpecName "kube-api-access-l78ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.832053 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9d5k\" (UniqueName: \"kubernetes.io/projected/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-kube-api-access-k9d5k\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.832086 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dd8bba3-489f-442a-97a4-47d2ebffdc9b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.832100 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l78ln\" (UniqueName: \"kubernetes.io/projected/e733dc01-8d9d-4a0b-b2e7-e18147562e11-kube-api-access-l78ln\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:46 crc kubenswrapper[4799]: I1124 07:06:46.832109 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpfw6\" (UniqueName: \"kubernetes.io/projected/5cde9df8-eba6-48ec-bc68-ed1fae49df81-kube-api-access-xpfw6\") on node \"crc\" DevicePath \"\"" Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.029913 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-722e-account-create-5cz9z" event={"ID":"3dd8bba3-489f-442a-97a4-47d2ebffdc9b","Type":"ContainerDied","Data":"c483e2cb2fdc62e52cbf180ed20c8ea4ec9f9d4de0c4cc2e42f7922b74507136"} Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.029964 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c483e2cb2fdc62e52cbf180ed20c8ea4ec9f9d4de0c4cc2e42f7922b74507136" Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.030033 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-722e-account-create-5cz9z" Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.035453 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b","Type":"ContainerStarted","Data":"62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1"} Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.037045 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-25d6-account-create-r6lpd" event={"ID":"5cde9df8-eba6-48ec-bc68-ed1fae49df81","Type":"ContainerDied","Data":"27fca7539965822ffb52d2fd4c09236b719b358427ec02abf084153067edd29a"} Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.037086 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27fca7539965822ffb52d2fd4c09236b719b358427ec02abf084153067edd29a" Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.037191 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-25d6-account-create-r6lpd" Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.043069 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-00c8-account-create-w4gf7" Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.043060 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-00c8-account-create-w4gf7" event={"ID":"04ddaef5-8849-4fa9-a0b9-f5e009dff420","Type":"ContainerDied","Data":"62cafdd129bdd495bc91cf610c9b7de9678a65a3adcfc244975b53cf904d5026"} Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.043200 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62cafdd129bdd495bc91cf610c9b7de9678a65a3adcfc244975b53cf904d5026" Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.045579 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z877t" Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.045648 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z877t" event={"ID":"558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed","Type":"ContainerDied","Data":"476830d22857f9fe260687edb887cfa22bb84f9b6dcbd2b10787126cad9ffa8c"} Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.045682 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="476830d22857f9fe260687edb887cfa22bb84f9b6dcbd2b10787126cad9ffa8c" Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.056530 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-mkgcm" event={"ID":"e733dc01-8d9d-4a0b-b2e7-e18147562e11","Type":"ContainerDied","Data":"ca3a4995cac84e1f8ca5ef118d5b97f96e61afc52a31a911b4f76f0881c200c8"} Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.056575 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca3a4995cac84e1f8ca5ef118d5b97f96e61afc52a31a911b4f76f0881c200c8" Nov 24 07:06:47 crc kubenswrapper[4799]: I1124 07:06:47.056582 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-mkgcm" Nov 24 07:06:48 crc kubenswrapper[4799]: I1124 07:06:48.069701 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b","Type":"ContainerStarted","Data":"613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53"} Nov 24 07:06:49 crc kubenswrapper[4799]: I1124 07:06:49.083528 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b","Type":"ContainerStarted","Data":"d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729"} Nov 24 07:06:49 crc kubenswrapper[4799]: I1124 07:06:49.083695 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="ceilometer-central-agent" containerID="cri-o://b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68" gracePeriod=30 Nov 24 07:06:49 crc kubenswrapper[4799]: I1124 07:06:49.083992 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:06:49 crc kubenswrapper[4799]: I1124 07:06:49.083990 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="proxy-httpd" containerID="cri-o://d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729" gracePeriod=30 Nov 24 07:06:49 crc kubenswrapper[4799]: I1124 07:06:49.084031 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="ceilometer-notification-agent" containerID="cri-o://62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1" gracePeriod=30 Nov 24 07:06:49 crc kubenswrapper[4799]: I1124 07:06:49.084041 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="sg-core" containerID="cri-o://613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53" gracePeriod=30 Nov 24 07:06:49 crc kubenswrapper[4799]: I1124 07:06:49.109372 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.651560408 podStartE2EDuration="6.109352502s" podCreationTimestamp="2025-11-24 07:06:43 +0000 UTC" firstStartedPulling="2025-11-24 07:06:44.260925449 +0000 UTC m=+1149.916907933" lastFinishedPulling="2025-11-24 07:06:48.718717553 +0000 UTC m=+1154.374700027" observedRunningTime="2025-11-24 07:06:49.104911637 +0000 UTC m=+1154.760894111" watchObservedRunningTime="2025-11-24 07:06:49.109352502 +0000 UTC m=+1154.765334976" Nov 24 07:06:50 crc kubenswrapper[4799]: I1124 07:06:50.100120 4799 generic.go:334] "Generic (PLEG): container finished" podID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerID="613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53" exitCode=2 Nov 24 07:06:50 crc kubenswrapper[4799]: I1124 07:06:50.100426 4799 generic.go:334] "Generic (PLEG): container finished" podID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerID="62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1" exitCode=0 Nov 24 07:06:50 crc kubenswrapper[4799]: I1124 07:06:50.100448 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b","Type":"ContainerDied","Data":"613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53"} Nov 24 07:06:50 crc kubenswrapper[4799]: I1124 07:06:50.100479 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b","Type":"ContainerDied","Data":"62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1"} Nov 24 07:06:50 crc kubenswrapper[4799]: I1124 07:06:50.400728 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:06:50 crc kubenswrapper[4799]: I1124 07:06:50.400784 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:06:50 crc kubenswrapper[4799]: I1124 07:06:50.400828 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:06:50 crc kubenswrapper[4799]: I1124 07:06:50.401614 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9829ac623e13e600dbb23e96d6ce6d0e396075d576f57770ed87771c2d8517cf"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:06:50 crc kubenswrapper[4799]: I1124 07:06:50.401725 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://9829ac623e13e600dbb23e96d6ce6d0e396075d576f57770ed87771c2d8517cf" gracePeriod=600 Nov 24 07:06:51 crc kubenswrapper[4799]: I1124 07:06:51.118092 4799 generic.go:334] "Generic (PLEG): container finished" podID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerID="b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68" exitCode=0 Nov 24 07:06:51 crc kubenswrapper[4799]: I1124 07:06:51.118177 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b","Type":"ContainerDied","Data":"b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68"} Nov 24 07:06:51 crc kubenswrapper[4799]: I1124 07:06:51.121916 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="9829ac623e13e600dbb23e96d6ce6d0e396075d576f57770ed87771c2d8517cf" exitCode=0 Nov 24 07:06:51 crc kubenswrapper[4799]: I1124 07:06:51.121975 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"9829ac623e13e600dbb23e96d6ce6d0e396075d576f57770ed87771c2d8517cf"} Nov 24 07:06:51 crc kubenswrapper[4799]: I1124 07:06:51.122009 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"5003777d46acb4b46479f05a79e8d64bd64dee47094406d45a1866a4471e6cd1"} Nov 24 07:06:51 crc kubenswrapper[4799]: I1124 07:06:51.122030 4799 scope.go:117] "RemoveContainer" containerID="985590a83c847e8770acbc79642e467b641116d42bb23156312a6fcbb0de12d9" Nov 24 07:06:51 crc kubenswrapper[4799]: I1124 07:06:51.457344 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:51 crc kubenswrapper[4799]: I1124 07:06:51.457999 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.618030 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8m4cw"] Nov 24 07:06:52 crc kubenswrapper[4799]: E1124 07:06:52.618696 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ddaef5-8849-4fa9-a0b9-f5e009dff420" containerName="mariadb-account-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.618713 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ddaef5-8849-4fa9-a0b9-f5e009dff420" containerName="mariadb-account-create" Nov 24 07:06:52 crc kubenswrapper[4799]: E1124 07:06:52.618726 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd8bba3-489f-442a-97a4-47d2ebffdc9b" containerName="mariadb-account-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.618735 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd8bba3-489f-442a-97a4-47d2ebffdc9b" containerName="mariadb-account-create" Nov 24 07:06:52 crc kubenswrapper[4799]: E1124 07:06:52.618765 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7130f73f-5578-4310-96f3-db076a954990" containerName="mariadb-database-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.618771 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7130f73f-5578-4310-96f3-db076a954990" containerName="mariadb-database-create" Nov 24 07:06:52 crc kubenswrapper[4799]: E1124 07:06:52.618785 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cde9df8-eba6-48ec-bc68-ed1fae49df81" containerName="mariadb-account-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.618791 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cde9df8-eba6-48ec-bc68-ed1fae49df81" containerName="mariadb-account-create" Nov 24 07:06:52 crc kubenswrapper[4799]: E1124 07:06:52.618804 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed" containerName="mariadb-database-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.618810 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed" containerName="mariadb-database-create" Nov 24 07:06:52 crc kubenswrapper[4799]: E1124 07:06:52.618829 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e733dc01-8d9d-4a0b-b2e7-e18147562e11" containerName="mariadb-database-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.618835 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e733dc01-8d9d-4a0b-b2e7-e18147562e11" containerName="mariadb-database-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.619023 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cde9df8-eba6-48ec-bc68-ed1fae49df81" containerName="mariadb-account-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.619038 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e733dc01-8d9d-4a0b-b2e7-e18147562e11" containerName="mariadb-database-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.619049 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7130f73f-5578-4310-96f3-db076a954990" containerName="mariadb-database-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.619058 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed" containerName="mariadb-database-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.619071 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd8bba3-489f-442a-97a4-47d2ebffdc9b" containerName="mariadb-account-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.619081 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="04ddaef5-8849-4fa9-a0b9-f5e009dff420" containerName="mariadb-account-create" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.619667 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.621483 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.621606 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.621973 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-s42wj" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.631910 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8m4cw"] Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.765427 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbn4t\" (UniqueName: \"kubernetes.io/projected/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-kube-api-access-qbn4t\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.765671 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-config-data\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.765704 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-scripts\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.765773 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.867597 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.867697 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbn4t\" (UniqueName: \"kubernetes.io/projected/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-kube-api-access-qbn4t\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.867788 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-config-data\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.867806 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-scripts\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.880915 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-config-data\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.885218 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-scripts\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.887685 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.890341 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbn4t\" (UniqueName: \"kubernetes.io/projected/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-kube-api-access-qbn4t\") pod \"nova-cell0-conductor-db-sync-8m4cw\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:52 crc kubenswrapper[4799]: I1124 07:06:52.949347 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:06:53 crc kubenswrapper[4799]: I1124 07:06:53.429393 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8m4cw"] Nov 24 07:06:53 crc kubenswrapper[4799]: W1124 07:06:53.434836 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bea0e41_5c29_42ed_9004_f72e9e7e22cd.slice/crio-99b8416bfb18b260b28f6ded0c90f823dd28b666c234fc924b2ddadcf92e1162 WatchSource:0}: Error finding container 99b8416bfb18b260b28f6ded0c90f823dd28b666c234fc924b2ddadcf92e1162: Status 404 returned error can't find the container with id 99b8416bfb18b260b28f6ded0c90f823dd28b666c234fc924b2ddadcf92e1162 Nov 24 07:06:54 crc kubenswrapper[4799]: I1124 07:06:54.161219 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8m4cw" event={"ID":"7bea0e41-5c29-42ed-9004-f72e9e7e22cd","Type":"ContainerStarted","Data":"99b8416bfb18b260b28f6ded0c90f823dd28b666c234fc924b2ddadcf92e1162"} Nov 24 07:06:57 crc kubenswrapper[4799]: I1124 07:06:57.478158 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:06:57 crc kubenswrapper[4799]: I1124 07:06:57.480213 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerName="glance-log" containerID="cri-o://4cb319180de5b546f429d89843adb706cbbabec96cc2e0ded3e2f44af9f83e16" gracePeriod=30 Nov 24 07:06:57 crc kubenswrapper[4799]: I1124 07:06:57.480280 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerName="glance-httpd" containerID="cri-o://f2093df971a210f5c37ee68ffd3cb3d71162fa62ac52dc98027d51155893ca23" gracePeriod=30 Nov 24 07:06:58 crc kubenswrapper[4799]: I1124 07:06:58.140894 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:06:58 crc kubenswrapper[4799]: I1124 07:06:58.141337 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerName="glance-log" containerID="cri-o://49b3ecf627707e4282cef4a93505d13d8a3d2349393427ee26e83ab6dc083bf4" gracePeriod=30 Nov 24 07:06:58 crc kubenswrapper[4799]: I1124 07:06:58.141432 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerName="glance-httpd" containerID="cri-o://ac1311b99af63466f8fdf3dde035f12ac5ad51bace15d1a69e8e29e4043627ac" gracePeriod=30 Nov 24 07:06:58 crc kubenswrapper[4799]: I1124 07:06:58.210215 4799 generic.go:334] "Generic (PLEG): container finished" podID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerID="4cb319180de5b546f429d89843adb706cbbabec96cc2e0ded3e2f44af9f83e16" exitCode=143 Nov 24 07:06:58 crc kubenswrapper[4799]: I1124 07:06:58.210260 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b6ec1b6c-dd0f-43e0-8894-d07fee864da5","Type":"ContainerDied","Data":"4cb319180de5b546f429d89843adb706cbbabec96cc2e0ded3e2f44af9f83e16"} Nov 24 07:06:59 crc kubenswrapper[4799]: I1124 07:06:59.240793 4799 generic.go:334] "Generic (PLEG): container finished" podID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerID="49b3ecf627707e4282cef4a93505d13d8a3d2349393427ee26e83ab6dc083bf4" exitCode=143 Nov 24 07:06:59 crc kubenswrapper[4799]: I1124 07:06:59.240866 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7d704f4-3e1b-4d87-807c-603b77175af5","Type":"ContainerDied","Data":"49b3ecf627707e4282cef4a93505d13d8a3d2349393427ee26e83ab6dc083bf4"} Nov 24 07:07:01 crc kubenswrapper[4799]: I1124 07:07:01.280900 4799 generic.go:334] "Generic (PLEG): container finished" podID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerID="f2093df971a210f5c37ee68ffd3cb3d71162fa62ac52dc98027d51155893ca23" exitCode=0 Nov 24 07:07:01 crc kubenswrapper[4799]: I1124 07:07:01.281135 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b6ec1b6c-dd0f-43e0-8894-d07fee864da5","Type":"ContainerDied","Data":"f2093df971a210f5c37ee68ffd3cb3d71162fa62ac52dc98027d51155893ca23"} Nov 24 07:07:01 crc kubenswrapper[4799]: I1124 07:07:01.976362 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.155:9292/healthcheck\": dial tcp 10.217.0.155:9292: connect: connection refused" Nov 24 07:07:01 crc kubenswrapper[4799]: I1124 07:07:01.978712 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.155:9292/healthcheck\": dial tcp 10.217.0.155:9292: connect: connection refused" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.006926 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.156:9292/healthcheck\": dial tcp 10.217.0.156:9292: connect: connection refused" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.007546 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.156:9292/healthcheck\": dial tcp 10.217.0.156:9292: connect: connection refused" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.255113 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.324314 4799 generic.go:334] "Generic (PLEG): container finished" podID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerID="ac1311b99af63466f8fdf3dde035f12ac5ad51bace15d1a69e8e29e4043627ac" exitCode=0 Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.324383 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7d704f4-3e1b-4d87-807c-603b77175af5","Type":"ContainerDied","Data":"ac1311b99af63466f8fdf3dde035f12ac5ad51bace15d1a69e8e29e4043627ac"} Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.329370 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b6ec1b6c-dd0f-43e0-8894-d07fee864da5","Type":"ContainerDied","Data":"acde285dbf829486a21a0e6e3a1134310a9ba803f0623ff5a73264f5b41cd2b9"} Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.329411 4799 scope.go:117] "RemoveContainer" containerID="f2093df971a210f5c37ee68ffd3cb3d71162fa62ac52dc98027d51155893ca23" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.329527 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.333000 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8m4cw" event={"ID":"7bea0e41-5c29-42ed-9004-f72e9e7e22cd","Type":"ContainerStarted","Data":"c138c8c4a165b11c836e1ac6661a13e3864e56da8aff5dba84d7403bc256e1bf"} Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.360955 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.361012 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-httpd-run\") pod \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.361085 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-public-tls-certs\") pod \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.361117 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-combined-ca-bundle\") pod \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.361175 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7blll\" (UniqueName: \"kubernetes.io/projected/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-kube-api-access-7blll\") pod \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.361567 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-logs\") pod \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.361690 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-scripts\") pod \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.361726 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-config-data\") pod \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\" (UID: \"b6ec1b6c-dd0f-43e0-8894-d07fee864da5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.363801 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b6ec1b6c-dd0f-43e0-8894-d07fee864da5" (UID: "b6ec1b6c-dd0f-43e0-8894-d07fee864da5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.364331 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-8m4cw" podStartSLOduration=1.8840356379999998 podStartE2EDuration="10.364313013s" podCreationTimestamp="2025-11-24 07:06:52 +0000 UTC" firstStartedPulling="2025-11-24 07:06:53.438080453 +0000 UTC m=+1159.094062937" lastFinishedPulling="2025-11-24 07:07:01.918357838 +0000 UTC m=+1167.574340312" observedRunningTime="2025-11-24 07:07:02.360405363 +0000 UTC m=+1168.016387837" watchObservedRunningTime="2025-11-24 07:07:02.364313013 +0000 UTC m=+1168.020295477" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.365834 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-logs" (OuterVolumeSpecName: "logs") pod "b6ec1b6c-dd0f-43e0-8894-d07fee864da5" (UID: "b6ec1b6c-dd0f-43e0-8894-d07fee864da5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.372805 4799 scope.go:117] "RemoveContainer" containerID="4cb319180de5b546f429d89843adb706cbbabec96cc2e0ded3e2f44af9f83e16" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.373478 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-scripts" (OuterVolumeSpecName: "scripts") pod "b6ec1b6c-dd0f-43e0-8894-d07fee864da5" (UID: "b6ec1b6c-dd0f-43e0-8894-d07fee864da5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.374007 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "b6ec1b6c-dd0f-43e0-8894-d07fee864da5" (UID: "b6ec1b6c-dd0f-43e0-8894-d07fee864da5"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.376459 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-kube-api-access-7blll" (OuterVolumeSpecName: "kube-api-access-7blll") pod "b6ec1b6c-dd0f-43e0-8894-d07fee864da5" (UID: "b6ec1b6c-dd0f-43e0-8894-d07fee864da5"). InnerVolumeSpecName "kube-api-access-7blll". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.406563 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.422911 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6ec1b6c-dd0f-43e0-8894-d07fee864da5" (UID: "b6ec1b6c-dd0f-43e0-8894-d07fee864da5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.447393 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-config-data" (OuterVolumeSpecName: "config-data") pod "b6ec1b6c-dd0f-43e0-8894-d07fee864da5" (UID: "b6ec1b6c-dd0f-43e0-8894-d07fee864da5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.466831 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.466879 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7blll\" (UniqueName: \"kubernetes.io/projected/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-kube-api-access-7blll\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.466893 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.466904 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.466916 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.466935 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.466944 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.493814 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b6ec1b6c-dd0f-43e0-8894-d07fee864da5" (UID: "b6ec1b6c-dd0f-43e0-8894-d07fee864da5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.503891 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.568453 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-internal-tls-certs\") pod \"f7d704f4-3e1b-4d87-807c-603b77175af5\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.568812 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-httpd-run\") pod \"f7d704f4-3e1b-4d87-807c-603b77175af5\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.568902 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"f7d704f4-3e1b-4d87-807c-603b77175af5\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.568956 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-combined-ca-bundle\") pod \"f7d704f4-3e1b-4d87-807c-603b77175af5\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.569008 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glvxs\" (UniqueName: \"kubernetes.io/projected/f7d704f4-3e1b-4d87-807c-603b77175af5-kube-api-access-glvxs\") pod \"f7d704f4-3e1b-4d87-807c-603b77175af5\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.569075 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-scripts\") pod \"f7d704f4-3e1b-4d87-807c-603b77175af5\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.569103 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-config-data\") pod \"f7d704f4-3e1b-4d87-807c-603b77175af5\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.569457 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f7d704f4-3e1b-4d87-807c-603b77175af5" (UID: "f7d704f4-3e1b-4d87-807c-603b77175af5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.569677 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-logs\") pod \"f7d704f4-3e1b-4d87-807c-603b77175af5\" (UID: \"f7d704f4-3e1b-4d87-807c-603b77175af5\") " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.570243 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.570267 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.570276 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6ec1b6c-dd0f-43e0-8894-d07fee864da5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.570335 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-logs" (OuterVolumeSpecName: "logs") pod "f7d704f4-3e1b-4d87-807c-603b77175af5" (UID: "f7d704f4-3e1b-4d87-807c-603b77175af5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.575290 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "f7d704f4-3e1b-4d87-807c-603b77175af5" (UID: "f7d704f4-3e1b-4d87-807c-603b77175af5"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.580201 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7d704f4-3e1b-4d87-807c-603b77175af5-kube-api-access-glvxs" (OuterVolumeSpecName: "kube-api-access-glvxs") pod "f7d704f4-3e1b-4d87-807c-603b77175af5" (UID: "f7d704f4-3e1b-4d87-807c-603b77175af5"). InnerVolumeSpecName "kube-api-access-glvxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.580211 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-scripts" (OuterVolumeSpecName: "scripts") pod "f7d704f4-3e1b-4d87-807c-603b77175af5" (UID: "f7d704f4-3e1b-4d87-807c-603b77175af5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.603656 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7d704f4-3e1b-4d87-807c-603b77175af5" (UID: "f7d704f4-3e1b-4d87-807c-603b77175af5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.624500 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-config-data" (OuterVolumeSpecName: "config-data") pod "f7d704f4-3e1b-4d87-807c-603b77175af5" (UID: "f7d704f4-3e1b-4d87-807c-603b77175af5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.638833 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f7d704f4-3e1b-4d87-807c-603b77175af5" (UID: "f7d704f4-3e1b-4d87-807c-603b77175af5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.667218 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.672527 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.672588 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.672602 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glvxs\" (UniqueName: \"kubernetes.io/projected/f7d704f4-3e1b-4d87-807c-603b77175af5-kube-api-access-glvxs\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.672612 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.672622 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.672631 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7d704f4-3e1b-4d87-807c-603b77175af5-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.672639 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7d704f4-3e1b-4d87-807c-603b77175af5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.678503 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.694711 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.703145 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:07:02 crc kubenswrapper[4799]: E1124 07:07:02.703542 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerName="glance-log" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.703568 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerName="glance-log" Nov 24 07:07:02 crc kubenswrapper[4799]: E1124 07:07:02.703593 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerName="glance-httpd" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.703602 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerName="glance-httpd" Nov 24 07:07:02 crc kubenswrapper[4799]: E1124 07:07:02.703622 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerName="glance-httpd" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.703630 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerName="glance-httpd" Nov 24 07:07:02 crc kubenswrapper[4799]: E1124 07:07:02.703642 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerName="glance-log" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.703648 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerName="glance-log" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.703839 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerName="glance-log" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.703913 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerName="glance-log" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.703925 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7d704f4-3e1b-4d87-807c-603b77175af5" containerName="glance-httpd" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.703933 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" containerName="glance-httpd" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.704867 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.707929 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.708176 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.728087 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.774580 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-logs\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.774656 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-scripts\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.774707 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.774756 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzcck\" (UniqueName: \"kubernetes.io/projected/e9a036d8-ce56-4654-a782-15cd18730378-kube-api-access-kzcck\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.774777 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.774799 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.774862 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.774887 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-config-data\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.774948 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.876785 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-scripts\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.876893 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.876942 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzcck\" (UniqueName: \"kubernetes.io/projected/e9a036d8-ce56-4654-a782-15cd18730378-kube-api-access-kzcck\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.876961 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.876983 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.876999 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.877400 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-config-data\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.877580 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-logs\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.877611 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.877969 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-logs\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.878416 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.882427 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.884234 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-scripts\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.886876 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.891586 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-config-data\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.899646 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzcck\" (UniqueName: \"kubernetes.io/projected/e9a036d8-ce56-4654-a782-15cd18730378-kube-api-access-kzcck\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:02 crc kubenswrapper[4799]: I1124 07:07:02.913529 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " pod="openstack/glance-default-external-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.019207 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.348164 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.348157 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7d704f4-3e1b-4d87-807c-603b77175af5","Type":"ContainerDied","Data":"6ee960a759929c1f2b370e54fa7cf51f11c31520e13144513846f8842770e57d"} Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.348677 4799 scope.go:117] "RemoveContainer" containerID="ac1311b99af63466f8fdf3dde035f12ac5ad51bace15d1a69e8e29e4043627ac" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.392100 4799 scope.go:117] "RemoveContainer" containerID="49b3ecf627707e4282cef4a93505d13d8a3d2349393427ee26e83ab6dc083bf4" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.419247 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.433246 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.448139 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.451126 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.467999 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.468194 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.489256 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.597981 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.598058 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.598093 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.598330 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.598704 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.598973 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.599050 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc7zx\" (UniqueName: \"kubernetes.io/projected/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-kube-api-access-fc7zx\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.599186 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.645301 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6ec1b6c-dd0f-43e0-8894-d07fee864da5" path="/var/lib/kubelet/pods/b6ec1b6c-dd0f-43e0-8894-d07fee864da5/volumes" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.677085 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7d704f4-3e1b-4d87-807c-603b77175af5" path="/var/lib/kubelet/pods/f7d704f4-3e1b-4d87-807c-603b77175af5/volumes" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.678543 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:07:03 crc kubenswrapper[4799]: W1124 07:07:03.687910 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9a036d8_ce56_4654_a782_15cd18730378.slice/crio-0b7a475aa5bac703f2bfb4027d1c6068f6d36bf5c155cd26fa56d69d8fe7c273 WatchSource:0}: Error finding container 0b7a475aa5bac703f2bfb4027d1c6068f6d36bf5c155cd26fa56d69d8fe7c273: Status 404 returned error can't find the container with id 0b7a475aa5bac703f2bfb4027d1c6068f6d36bf5c155cd26fa56d69d8fe7c273 Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.703604 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.703680 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc7zx\" (UniqueName: \"kubernetes.io/projected/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-kube-api-access-fc7zx\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.703721 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.703802 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.703822 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.703900 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.703950 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.704008 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.704605 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.704879 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.709074 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.717141 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.717439 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.717795 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.718400 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.734479 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc7zx\" (UniqueName: \"kubernetes.io/projected/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-kube-api-access-fc7zx\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.769476 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " pod="openstack/glance-default-internal-api-0" Nov 24 07:07:03 crc kubenswrapper[4799]: I1124 07:07:03.779218 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:04 crc kubenswrapper[4799]: I1124 07:07:04.362817 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e9a036d8-ce56-4654-a782-15cd18730378","Type":"ContainerStarted","Data":"0b7a475aa5bac703f2bfb4027d1c6068f6d36bf5c155cd26fa56d69d8fe7c273"} Nov 24 07:07:04 crc kubenswrapper[4799]: I1124 07:07:04.372471 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:07:04 crc kubenswrapper[4799]: W1124 07:07:04.377577 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b14e8a8_2e19_47fd_a440_ac8d0cafecaf.slice/crio-6256ad3dda3b8c8f22ca60c97059524d69b86c6cc089604645377e1368bb932f WatchSource:0}: Error finding container 6256ad3dda3b8c8f22ca60c97059524d69b86c6cc089604645377e1368bb932f: Status 404 returned error can't find the container with id 6256ad3dda3b8c8f22ca60c97059524d69b86c6cc089604645377e1368bb932f Nov 24 07:07:05 crc kubenswrapper[4799]: I1124 07:07:05.376594 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf","Type":"ContainerStarted","Data":"668ba3f136f6d97ba947f42f79e2823ce4845dbde12c05bc8fe71572551507f9"} Nov 24 07:07:05 crc kubenswrapper[4799]: I1124 07:07:05.377252 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf","Type":"ContainerStarted","Data":"6256ad3dda3b8c8f22ca60c97059524d69b86c6cc089604645377e1368bb932f"} Nov 24 07:07:05 crc kubenswrapper[4799]: I1124 07:07:05.379632 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e9a036d8-ce56-4654-a782-15cd18730378","Type":"ContainerStarted","Data":"e97f64b93e8c87fb3e9989776d9be9cfbdd54d15ade6362a154f53220314a180"} Nov 24 07:07:05 crc kubenswrapper[4799]: I1124 07:07:05.379708 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e9a036d8-ce56-4654-a782-15cd18730378","Type":"ContainerStarted","Data":"bb47a5f8acf503e4173095e108c8eaae5b42159d81d170336e69837e3fecd4b9"} Nov 24 07:07:05 crc kubenswrapper[4799]: I1124 07:07:05.412454 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.412435151 podStartE2EDuration="3.412435151s" podCreationTimestamp="2025-11-24 07:07:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:05.399351369 +0000 UTC m=+1171.055333863" watchObservedRunningTime="2025-11-24 07:07:05.412435151 +0000 UTC m=+1171.068417625" Nov 24 07:07:07 crc kubenswrapper[4799]: I1124 07:07:07.400554 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf","Type":"ContainerStarted","Data":"183c3c900b89e4d354f72e7ee6593a3ca4a44de57d248b7f61ac4959fe51c8ef"} Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.019832 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.020428 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.057758 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.081124 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.126576 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=10.126557217 podStartE2EDuration="10.126557217s" podCreationTimestamp="2025-11-24 07:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:07.426239506 +0000 UTC m=+1173.082221990" watchObservedRunningTime="2025-11-24 07:07:13.126557217 +0000 UTC m=+1178.782539691" Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.474052 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.474121 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.603830 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.782381 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.782458 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.813883 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:13 crc kubenswrapper[4799]: I1124 07:07:13.823567 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:14 crc kubenswrapper[4799]: I1124 07:07:14.487150 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:14 crc kubenswrapper[4799]: I1124 07:07:14.487559 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:15 crc kubenswrapper[4799]: I1124 07:07:15.375655 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 07:07:15 crc kubenswrapper[4799]: I1124 07:07:15.421157 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 07:07:15 crc kubenswrapper[4799]: I1124 07:07:15.501780 4799 generic.go:334] "Generic (PLEG): container finished" podID="7bea0e41-5c29-42ed-9004-f72e9e7e22cd" containerID="c138c8c4a165b11c836e1ac6661a13e3864e56da8aff5dba84d7403bc256e1bf" exitCode=0 Nov 24 07:07:15 crc kubenswrapper[4799]: I1124 07:07:15.501996 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8m4cw" event={"ID":"7bea0e41-5c29-42ed-9004-f72e9e7e22cd","Type":"ContainerDied","Data":"c138c8c4a165b11c836e1ac6661a13e3864e56da8aff5dba84d7403bc256e1bf"} Nov 24 07:07:16 crc kubenswrapper[4799]: I1124 07:07:16.561457 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:16 crc kubenswrapper[4799]: I1124 07:07:16.562218 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 07:07:16 crc kubenswrapper[4799]: I1124 07:07:16.568852 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 07:07:16 crc kubenswrapper[4799]: I1124 07:07:16.955308 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.027036 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-scripts\") pod \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.027135 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-config-data\") pod \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.027242 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbn4t\" (UniqueName: \"kubernetes.io/projected/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-kube-api-access-qbn4t\") pod \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.027371 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-combined-ca-bundle\") pod \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\" (UID: \"7bea0e41-5c29-42ed-9004-f72e9e7e22cd\") " Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.035712 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-kube-api-access-qbn4t" (OuterVolumeSpecName: "kube-api-access-qbn4t") pod "7bea0e41-5c29-42ed-9004-f72e9e7e22cd" (UID: "7bea0e41-5c29-42ed-9004-f72e9e7e22cd"). InnerVolumeSpecName "kube-api-access-qbn4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.048281 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-scripts" (OuterVolumeSpecName: "scripts") pod "7bea0e41-5c29-42ed-9004-f72e9e7e22cd" (UID: "7bea0e41-5c29-42ed-9004-f72e9e7e22cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.068190 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-config-data" (OuterVolumeSpecName: "config-data") pod "7bea0e41-5c29-42ed-9004-f72e9e7e22cd" (UID: "7bea0e41-5c29-42ed-9004-f72e9e7e22cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.084743 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bea0e41-5c29-42ed-9004-f72e9e7e22cd" (UID: "7bea0e41-5c29-42ed-9004-f72e9e7e22cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.131312 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.131384 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.131400 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.131414 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbn4t\" (UniqueName: \"kubernetes.io/projected/7bea0e41-5c29-42ed-9004-f72e9e7e22cd-kube-api-access-qbn4t\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.527207 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8m4cw" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.534227 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8m4cw" event={"ID":"7bea0e41-5c29-42ed-9004-f72e9e7e22cd","Type":"ContainerDied","Data":"99b8416bfb18b260b28f6ded0c90f823dd28b666c234fc924b2ddadcf92e1162"} Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.534297 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99b8416bfb18b260b28f6ded0c90f823dd28b666c234fc924b2ddadcf92e1162" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.640933 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 07:07:17 crc kubenswrapper[4799]: E1124 07:07:17.641341 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bea0e41-5c29-42ed-9004-f72e9e7e22cd" containerName="nova-cell0-conductor-db-sync" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.641358 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bea0e41-5c29-42ed-9004-f72e9e7e22cd" containerName="nova-cell0-conductor-db-sync" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.641597 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bea0e41-5c29-42ed-9004-f72e9e7e22cd" containerName="nova-cell0-conductor-db-sync" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.642429 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.647280 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.647494 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-s42wj" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.653603 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.743963 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7nqh\" (UniqueName: \"kubernetes.io/projected/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-kube-api-access-s7nqh\") pod \"nova-cell0-conductor-0\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.744014 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.744231 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.846575 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7nqh\" (UniqueName: \"kubernetes.io/projected/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-kube-api-access-s7nqh\") pod \"nova-cell0-conductor-0\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.846634 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.846713 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.850684 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.861756 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.865628 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7nqh\" (UniqueName: \"kubernetes.io/projected/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-kube-api-access-s7nqh\") pod \"nova-cell0-conductor-0\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:17 crc kubenswrapper[4799]: I1124 07:07:17.979236 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:18 crc kubenswrapper[4799]: I1124 07:07:18.469668 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 07:07:18 crc kubenswrapper[4799]: I1124 07:07:18.535764 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1","Type":"ContainerStarted","Data":"1d129c757ca1658503bc21c474818337e63f0f84ff4d13fb65c9e13277d1a15c"} Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.510488 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.559795 4799 generic.go:334] "Generic (PLEG): container finished" podID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerID="d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729" exitCode=137 Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.559878 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b","Type":"ContainerDied","Data":"d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729"} Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.559939 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.560045 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b","Type":"ContainerDied","Data":"cf4f4a234f688bdd7d36b9245197a806bfdda60c4553c7fd0012c9cc8cb32922"} Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.560081 4799 scope.go:117] "RemoveContainer" containerID="d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.566047 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1","Type":"ContainerStarted","Data":"8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05"} Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.566205 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.584568 4799 scope.go:117] "RemoveContainer" containerID="613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.586682 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-run-httpd\") pod \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.586770 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-config-data\") pod \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.586858 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jcmj\" (UniqueName: \"kubernetes.io/projected/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-kube-api-access-8jcmj\") pod \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.586897 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-sg-core-conf-yaml\") pod \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.587008 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-scripts\") pod \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.587102 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-combined-ca-bundle\") pod \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.587179 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-log-httpd\") pod \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\" (UID: \"f8c58e1a-1d91-4f16-843a-bd6f81a03d7b\") " Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.588075 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" (UID: "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.590204 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" (UID: "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.594772 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.5947483460000003 podStartE2EDuration="2.594748346s" podCreationTimestamp="2025-11-24 07:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:19.585629657 +0000 UTC m=+1185.241612131" watchObservedRunningTime="2025-11-24 07:07:19.594748346 +0000 UTC m=+1185.250730820" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.596200 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-kube-api-access-8jcmj" (OuterVolumeSpecName: "kube-api-access-8jcmj") pod "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" (UID: "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b"). InnerVolumeSpecName "kube-api-access-8jcmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.596800 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-scripts" (OuterVolumeSpecName: "scripts") pod "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" (UID: "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.616988 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" (UID: "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.655701 4799 scope.go:117] "RemoveContainer" containerID="62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.680391 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" (UID: "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.683814 4799 scope.go:117] "RemoveContainer" containerID="b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.690061 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jcmj\" (UniqueName: \"kubernetes.io/projected/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-kube-api-access-8jcmj\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.690115 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.690125 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.690136 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.690349 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.691239 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.692150 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-config-data" (OuterVolumeSpecName: "config-data") pod "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" (UID: "f8c58e1a-1d91-4f16-843a-bd6f81a03d7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.712333 4799 scope.go:117] "RemoveContainer" containerID="d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729" Nov 24 07:07:19 crc kubenswrapper[4799]: E1124 07:07:19.713061 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729\": container with ID starting with d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729 not found: ID does not exist" containerID="d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.713126 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729"} err="failed to get container status \"d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729\": rpc error: code = NotFound desc = could not find container \"d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729\": container with ID starting with d3c204916969a2dc53615200a1b440c524ef70d2abd21328beeaaa14267a5729 not found: ID does not exist" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.713170 4799 scope.go:117] "RemoveContainer" containerID="613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53" Nov 24 07:07:19 crc kubenswrapper[4799]: E1124 07:07:19.713915 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53\": container with ID starting with 613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53 not found: ID does not exist" containerID="613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.713975 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53"} err="failed to get container status \"613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53\": rpc error: code = NotFound desc = could not find container \"613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53\": container with ID starting with 613a513eb89d2e505414b205ab55bbc37ef9da450d3edab170302c3b3c977f53 not found: ID does not exist" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.714015 4799 scope.go:117] "RemoveContainer" containerID="62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1" Nov 24 07:07:19 crc kubenswrapper[4799]: E1124 07:07:19.714573 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1\": container with ID starting with 62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1 not found: ID does not exist" containerID="62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.714605 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1"} err="failed to get container status \"62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1\": rpc error: code = NotFound desc = could not find container \"62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1\": container with ID starting with 62a4f61d876655e8ffd07b3f05eb8940c5d707ad5897ecd367cde66d8ed969a1 not found: ID does not exist" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.714626 4799 scope.go:117] "RemoveContainer" containerID="b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68" Nov 24 07:07:19 crc kubenswrapper[4799]: E1124 07:07:19.715488 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68\": container with ID starting with b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68 not found: ID does not exist" containerID="b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.716230 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68"} err="failed to get container status \"b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68\": rpc error: code = NotFound desc = could not find container \"b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68\": container with ID starting with b401ad09e11aec731d73b1808c8f02ed450f88da0a6102339c00e2dc23d76e68 not found: ID does not exist" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.793111 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.946657 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.963755 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.989576 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:07:19 crc kubenswrapper[4799]: E1124 07:07:19.990139 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="proxy-httpd" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.990163 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="proxy-httpd" Nov 24 07:07:19 crc kubenswrapper[4799]: E1124 07:07:19.990178 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="ceilometer-notification-agent" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.990187 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="ceilometer-notification-agent" Nov 24 07:07:19 crc kubenswrapper[4799]: E1124 07:07:19.990210 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="sg-core" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.990218 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="sg-core" Nov 24 07:07:19 crc kubenswrapper[4799]: E1124 07:07:19.990247 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="ceilometer-central-agent" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.990255 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="ceilometer-central-agent" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.990497 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="ceilometer-notification-agent" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.990514 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="proxy-httpd" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.990528 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="sg-core" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.990560 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" containerName="ceilometer-central-agent" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.992766 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.995486 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:07:19 crc kubenswrapper[4799]: I1124 07:07:19.996491 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.005182 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.100655 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.100714 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-config-data\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.100776 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv7j2\" (UniqueName: \"kubernetes.io/projected/2119f5af-764e-4301-9b73-e44b3771c62a-kube-api-access-tv7j2\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.100813 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-scripts\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.100874 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.100937 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-log-httpd\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.100984 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-run-httpd\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.202652 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-log-httpd\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.202717 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-run-httpd\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.202784 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.202806 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-config-data\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.202867 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv7j2\" (UniqueName: \"kubernetes.io/projected/2119f5af-764e-4301-9b73-e44b3771c62a-kube-api-access-tv7j2\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.202900 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-scripts\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.202938 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.204421 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-run-httpd\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.204484 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-log-httpd\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.209290 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-scripts\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.210348 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-config-data\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.211304 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.222624 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.230159 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv7j2\" (UniqueName: \"kubernetes.io/projected/2119f5af-764e-4301-9b73-e44b3771c62a-kube-api-access-tv7j2\") pod \"ceilometer-0\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.312717 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:07:20 crc kubenswrapper[4799]: I1124 07:07:20.800801 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:07:20 crc kubenswrapper[4799]: W1124 07:07:20.801871 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2119f5af_764e_4301_9b73_e44b3771c62a.slice/crio-81a47571abcba2e0656a1ce37b514272f6eb33b0b6fba845d04cf613247158f1 WatchSource:0}: Error finding container 81a47571abcba2e0656a1ce37b514272f6eb33b0b6fba845d04cf613247158f1: Status 404 returned error can't find the container with id 81a47571abcba2e0656a1ce37b514272f6eb33b0b6fba845d04cf613247158f1 Nov 24 07:07:21 crc kubenswrapper[4799]: I1124 07:07:21.589493 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2119f5af-764e-4301-9b73-e44b3771c62a","Type":"ContainerStarted","Data":"f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663"} Nov 24 07:07:21 crc kubenswrapper[4799]: I1124 07:07:21.589544 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2119f5af-764e-4301-9b73-e44b3771c62a","Type":"ContainerStarted","Data":"81a47571abcba2e0656a1ce37b514272f6eb33b0b6fba845d04cf613247158f1"} Nov 24 07:07:21 crc kubenswrapper[4799]: I1124 07:07:21.639615 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8c58e1a-1d91-4f16-843a-bd6f81a03d7b" path="/var/lib/kubelet/pods/f8c58e1a-1d91-4f16-843a-bd6f81a03d7b/volumes" Nov 24 07:07:22 crc kubenswrapper[4799]: I1124 07:07:22.611019 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2119f5af-764e-4301-9b73-e44b3771c62a","Type":"ContainerStarted","Data":"f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31"} Nov 24 07:07:23 crc kubenswrapper[4799]: I1124 07:07:23.654999 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2119f5af-764e-4301-9b73-e44b3771c62a","Type":"ContainerStarted","Data":"ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786"} Nov 24 07:07:24 crc kubenswrapper[4799]: I1124 07:07:24.666332 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2119f5af-764e-4301-9b73-e44b3771c62a","Type":"ContainerStarted","Data":"076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6"} Nov 24 07:07:24 crc kubenswrapper[4799]: I1124 07:07:24.666921 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:07:24 crc kubenswrapper[4799]: I1124 07:07:24.695314 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.44776051 podStartE2EDuration="5.695288636s" podCreationTimestamp="2025-11-24 07:07:19 +0000 UTC" firstStartedPulling="2025-11-24 07:07:20.804812605 +0000 UTC m=+1186.460795089" lastFinishedPulling="2025-11-24 07:07:24.052340741 +0000 UTC m=+1189.708323215" observedRunningTime="2025-11-24 07:07:24.693934438 +0000 UTC m=+1190.349916952" watchObservedRunningTime="2025-11-24 07:07:24.695288636 +0000 UTC m=+1190.351271110" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.011224 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.572770 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-fj4rj"] Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.574143 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.576609 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.577905 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.592084 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fj4rj"] Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.663095 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.663217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-config-data\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.663262 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwk9l\" (UniqueName: \"kubernetes.io/projected/959e83f3-69f2-433b-a326-242ec1c4b226-kube-api-access-dwk9l\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.663299 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-scripts\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.764986 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.765058 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-config-data\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.765093 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwk9l\" (UniqueName: \"kubernetes.io/projected/959e83f3-69f2-433b-a326-242ec1c4b226-kube-api-access-dwk9l\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.765133 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-scripts\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.771100 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-config-data\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.771101 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.773454 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-scripts\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.792406 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwk9l\" (UniqueName: \"kubernetes.io/projected/959e83f3-69f2-433b-a326-242ec1c4b226-kube-api-access-dwk9l\") pod \"nova-cell0-cell-mapping-fj4rj\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.829876 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.838189 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.840684 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.874298 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.874406 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.874482 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt2sf\" (UniqueName: \"kubernetes.io/projected/4c989b53-8318-4b5d-b893-20c0d1e98cd4-kube-api-access-wt2sf\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.885601 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.898140 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.898789 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.902177 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.906867 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.914271 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.978320 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.978396 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svgvh\" (UniqueName: \"kubernetes.io/projected/002e276b-143e-4062-8530-ffaec1bdc2b8-kube-api-access-svgvh\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.978438 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.978483 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-config-data\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.978543 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt2sf\" (UniqueName: \"kubernetes.io/projected/4c989b53-8318-4b5d-b893-20c0d1e98cd4-kube-api-access-wt2sf\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.978579 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/002e276b-143e-4062-8530-ffaec1bdc2b8-logs\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.978639 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:28 crc kubenswrapper[4799]: I1124 07:07:28.995150 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.000616 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.010246 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt2sf\" (UniqueName: \"kubernetes.io/projected/4c989b53-8318-4b5d-b893-20c0d1e98cd4-kube-api-access-wt2sf\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.058266 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-bk2hb"] Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.062744 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.079941 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.080212 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-config\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.080376 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.080490 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svgvh\" (UniqueName: \"kubernetes.io/projected/002e276b-143e-4062-8530-ffaec1bdc2b8-kube-api-access-svgvh\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.080572 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-swift-storage-0\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.080649 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.080721 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t845r\" (UniqueName: \"kubernetes.io/projected/f60aa6ed-3836-4943-b93f-9b4f221cb82e-kube-api-access-t845r\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.080812 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-config-data\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.080921 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-svc\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.081046 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/002e276b-143e-4062-8530-ffaec1bdc2b8-logs\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.081624 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/002e276b-143e-4062-8530-ffaec1bdc2b8-logs\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.087549 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.087614 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-bk2hb"] Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.096363 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.098797 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-config-data\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.101818 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.105830 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.110790 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.130739 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.157252 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svgvh\" (UniqueName: \"kubernetes.io/projected/002e276b-143e-4062-8530-ffaec1bdc2b8-kube-api-access-svgvh\") pod \"nova-metadata-0\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " pod="openstack/nova-metadata-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.196711 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-swift-storage-0\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.196812 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t845r\" (UniqueName: \"kubernetes.io/projected/f60aa6ed-3836-4943-b93f-9b4f221cb82e-kube-api-access-t845r\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.196967 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-svc\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.197213 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.197381 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-config\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.197436 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.215743 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.216503 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-swift-storage-0\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.232193 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.236215 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-config\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.238679 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-svc\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.252047 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t845r\" (UniqueName: \"kubernetes.io/projected/f60aa6ed-3836-4943-b93f-9b4f221cb82e-kube-api-access-t845r\") pod \"dnsmasq-dns-5dd7c4987f-bk2hb\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.294102 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.296903 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.304053 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.305020 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx792\" (UniqueName: \"kubernetes.io/projected/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-kube-api-access-sx792\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.305100 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-logs\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.305200 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-config-data\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.305280 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.306772 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.407442 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.407975 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.408112 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx792\" (UniqueName: \"kubernetes.io/projected/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-kube-api-access-sx792\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.408216 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-logs\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.408293 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkz8h\" (UniqueName: \"kubernetes.io/projected/5a696130-3b34-4a3b-8ab9-677077939d0f-kube-api-access-xkz8h\") pod \"nova-scheduler-0\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.408366 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-config-data\") pod \"nova-scheduler-0\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.408504 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-config-data\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.410220 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-logs\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.418458 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-config-data\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.428185 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.438556 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.444592 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx792\" (UniqueName: \"kubernetes.io/projected/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-kube-api-access-sx792\") pod \"nova-api-0\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.499458 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.511623 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkz8h\" (UniqueName: \"kubernetes.io/projected/5a696130-3b34-4a3b-8ab9-677077939d0f-kube-api-access-xkz8h\") pod \"nova-scheduler-0\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.511669 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-config-data\") pod \"nova-scheduler-0\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.511752 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.515322 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-config-data\") pod \"nova-scheduler-0\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.517663 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.534585 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkz8h\" (UniqueName: \"kubernetes.io/projected/5a696130-3b34-4a3b-8ab9-677077939d0f-kube-api-access-xkz8h\") pod \"nova-scheduler-0\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.535055 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.544820 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fj4rj"] Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.625989 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.718327 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-k9t68"] Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.720013 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.723357 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.723702 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.732529 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-k9t68"] Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.749494 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.750289 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fj4rj" event={"ID":"959e83f3-69f2-433b-a326-242ec1c4b226","Type":"ContainerStarted","Data":"6a03dbaea59b59d37f24b9630ae4e68d05d1ae5f73b9763c1d53867e08b1ae63"} Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.817703 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.818105 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-scripts\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.818162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmddr\" (UniqueName: \"kubernetes.io/projected/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-kube-api-access-hmddr\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.818244 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-config-data\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.920344 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmddr\" (UniqueName: \"kubernetes.io/projected/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-kube-api-access-hmddr\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.920490 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-config-data\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.920543 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.920631 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-scripts\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.925962 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-config-data\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.928544 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.931838 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-scripts\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.950278 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmddr\" (UniqueName: \"kubernetes.io/projected/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-kube-api-access-hmddr\") pod \"nova-cell1-conductor-db-sync-k9t68\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.954330 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:29 crc kubenswrapper[4799]: I1124 07:07:29.986421 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-bk2hb"] Nov 24 07:07:30 crc kubenswrapper[4799]: W1124 07:07:30.032063 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf60aa6ed_3836_4943_b93f_9b4f221cb82e.slice/crio-1730fa91e1072e574816dacd5cc12229d9a79ab889b1d0574c7b90beca763cb5 WatchSource:0}: Error finding container 1730fa91e1072e574816dacd5cc12229d9a79ab889b1d0574c7b90beca763cb5: Status 404 returned error can't find the container with id 1730fa91e1072e574816dacd5cc12229d9a79ab889b1d0574c7b90beca763cb5 Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.057551 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.301029 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.398332 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.601296 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-k9t68"] Nov 24 07:07:30 crc kubenswrapper[4799]: W1124 07:07:30.607242 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod650550d4_96f8_46dd_9117_d4f9c2cdcd9d.slice/crio-59a7d6d0f4e4fa374070373f9908120dbcf22bb6997cd55465802dcff6d1a645 WatchSource:0}: Error finding container 59a7d6d0f4e4fa374070373f9908120dbcf22bb6997cd55465802dcff6d1a645: Status 404 returned error can't find the container with id 59a7d6d0f4e4fa374070373f9908120dbcf22bb6997cd55465802dcff6d1a645 Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.778893 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a696130-3b34-4a3b-8ab9-677077939d0f","Type":"ContainerStarted","Data":"cf3af3c3025f3af56dfb33fffc7204978605c8c7d5a0d68f6749e9d4b731993f"} Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.786115 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-k9t68" event={"ID":"650550d4-96f8-46dd-9117-d4f9c2cdcd9d","Type":"ContainerStarted","Data":"59a7d6d0f4e4fa374070373f9908120dbcf22bb6997cd55465802dcff6d1a645"} Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.799582 4799 generic.go:334] "Generic (PLEG): container finished" podID="f60aa6ed-3836-4943-b93f-9b4f221cb82e" containerID="b238dc5989e8a0ab0a49951d5190f429b15a6d6522ab9d6c0b546c98d7071513" exitCode=0 Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.799979 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" event={"ID":"f60aa6ed-3836-4943-b93f-9b4f221cb82e","Type":"ContainerDied","Data":"b238dc5989e8a0ab0a49951d5190f429b15a6d6522ab9d6c0b546c98d7071513"} Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.800024 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" event={"ID":"f60aa6ed-3836-4943-b93f-9b4f221cb82e","Type":"ContainerStarted","Data":"1730fa91e1072e574816dacd5cc12229d9a79ab889b1d0574c7b90beca763cb5"} Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.821962 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4c989b53-8318-4b5d-b893-20c0d1e98cd4","Type":"ContainerStarted","Data":"145996eed3db13e8d38ac0c13cf4336848c1f7bb23b558439ebbaa32800e3113"} Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.826720 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fj4rj" event={"ID":"959e83f3-69f2-433b-a326-242ec1c4b226","Type":"ContainerStarted","Data":"88399e1c222bc6dd7a643d5db0cd6a4e6b60cf5019ca0bbc1ca526de8ed139d3"} Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.835677 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8","Type":"ContainerStarted","Data":"af5f7cd771c629cbcd61bb98a6da710202ea0378475baacf83fb3a7d21b49810"} Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.851104 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"002e276b-143e-4062-8530-ffaec1bdc2b8","Type":"ContainerStarted","Data":"d4e6f97f39b33eb1f799ff6126af14bbee59cd9687555acd992d6a6472b1e48f"} Nov 24 07:07:30 crc kubenswrapper[4799]: I1124 07:07:30.856975 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-fj4rj" podStartSLOduration=2.856952967 podStartE2EDuration="2.856952967s" podCreationTimestamp="2025-11-24 07:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:30.846734517 +0000 UTC m=+1196.502716991" watchObservedRunningTime="2025-11-24 07:07:30.856952967 +0000 UTC m=+1196.512935441" Nov 24 07:07:31 crc kubenswrapper[4799]: I1124 07:07:31.893208 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" event={"ID":"f60aa6ed-3836-4943-b93f-9b4f221cb82e","Type":"ContainerStarted","Data":"c1b436269325ee06b16c543789a8532f1ca29f89d39ac77a6252f8a6b10370f3"} Nov 24 07:07:31 crc kubenswrapper[4799]: I1124 07:07:31.893625 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:31 crc kubenswrapper[4799]: I1124 07:07:31.908682 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-k9t68" event={"ID":"650550d4-96f8-46dd-9117-d4f9c2cdcd9d","Type":"ContainerStarted","Data":"c4a15beff8279fb5630bb3a06f94fc601d98cba69dc34d057239038e1cbe6e62"} Nov 24 07:07:31 crc kubenswrapper[4799]: I1124 07:07:31.915776 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" podStartSLOduration=3.915755033 podStartE2EDuration="3.915755033s" podCreationTimestamp="2025-11-24 07:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:31.91317775 +0000 UTC m=+1197.569160224" watchObservedRunningTime="2025-11-24 07:07:31.915755033 +0000 UTC m=+1197.571737507" Nov 24 07:07:31 crc kubenswrapper[4799]: I1124 07:07:31.954031 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-k9t68" podStartSLOduration=2.954009799 podStartE2EDuration="2.954009799s" podCreationTimestamp="2025-11-24 07:07:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:31.935681359 +0000 UTC m=+1197.591663823" watchObservedRunningTime="2025-11-24 07:07:31.954009799 +0000 UTC m=+1197.609992273" Nov 24 07:07:32 crc kubenswrapper[4799]: I1124 07:07:32.624169 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:32 crc kubenswrapper[4799]: I1124 07:07:32.641379 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:07:33 crc kubenswrapper[4799]: I1124 07:07:33.922253 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4c989b53-8318-4b5d-b893-20c0d1e98cd4","Type":"ContainerStarted","Data":"28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b"} Nov 24 07:07:33 crc kubenswrapper[4799]: I1124 07:07:33.922313 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="4c989b53-8318-4b5d-b893-20c0d1e98cd4" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b" gracePeriod=30 Nov 24 07:07:33 crc kubenswrapper[4799]: I1124 07:07:33.925052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8","Type":"ContainerStarted","Data":"92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e"} Nov 24 07:07:33 crc kubenswrapper[4799]: I1124 07:07:33.930656 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"002e276b-143e-4062-8530-ffaec1bdc2b8","Type":"ContainerStarted","Data":"4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc"} Nov 24 07:07:33 crc kubenswrapper[4799]: I1124 07:07:33.941962 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a696130-3b34-4a3b-8ab9-677077939d0f","Type":"ContainerStarted","Data":"4f037689b65a1da575275f13da9f49f1ca9ea45db68ab593237cee43fc871a9b"} Nov 24 07:07:33 crc kubenswrapper[4799]: I1124 07:07:33.947906 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.319372501 podStartE2EDuration="5.947884969s" podCreationTimestamp="2025-11-24 07:07:28 +0000 UTC" firstStartedPulling="2025-11-24 07:07:29.786216192 +0000 UTC m=+1195.442198666" lastFinishedPulling="2025-11-24 07:07:33.41472866 +0000 UTC m=+1199.070711134" observedRunningTime="2025-11-24 07:07:33.939684227 +0000 UTC m=+1199.595666701" watchObservedRunningTime="2025-11-24 07:07:33.947884969 +0000 UTC m=+1199.603867443" Nov 24 07:07:33 crc kubenswrapper[4799]: I1124 07:07:33.967032 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.008089796 podStartE2EDuration="4.967014962s" podCreationTimestamp="2025-11-24 07:07:29 +0000 UTC" firstStartedPulling="2025-11-24 07:07:30.45637089 +0000 UTC m=+1196.112353364" lastFinishedPulling="2025-11-24 07:07:33.415296046 +0000 UTC m=+1199.071278530" observedRunningTime="2025-11-24 07:07:33.957280816 +0000 UTC m=+1199.613263290" watchObservedRunningTime="2025-11-24 07:07:33.967014962 +0000 UTC m=+1199.622997436" Nov 24 07:07:34 crc kubenswrapper[4799]: I1124 07:07:34.107440 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:07:34 crc kubenswrapper[4799]: I1124 07:07:34.627363 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 07:07:34 crc kubenswrapper[4799]: I1124 07:07:34.957020 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8","Type":"ContainerStarted","Data":"f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04"} Nov 24 07:07:34 crc kubenswrapper[4799]: I1124 07:07:34.959138 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"002e276b-143e-4062-8530-ffaec1bdc2b8","Type":"ContainerStarted","Data":"319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92"} Nov 24 07:07:34 crc kubenswrapper[4799]: I1124 07:07:34.959319 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="002e276b-143e-4062-8530-ffaec1bdc2b8" containerName="nova-metadata-metadata" containerID="cri-o://319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92" gracePeriod=30 Nov 24 07:07:34 crc kubenswrapper[4799]: I1124 07:07:34.959310 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="002e276b-143e-4062-8530-ffaec1bdc2b8" containerName="nova-metadata-log" containerID="cri-o://4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc" gracePeriod=30 Nov 24 07:07:34 crc kubenswrapper[4799]: I1124 07:07:34.993323 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.920102847 podStartE2EDuration="5.993304626s" podCreationTimestamp="2025-11-24 07:07:29 +0000 UTC" firstStartedPulling="2025-11-24 07:07:30.350593258 +0000 UTC m=+1196.006575732" lastFinishedPulling="2025-11-24 07:07:33.423795037 +0000 UTC m=+1199.079777511" observedRunningTime="2025-11-24 07:07:34.987876162 +0000 UTC m=+1200.643858636" watchObservedRunningTime="2025-11-24 07:07:34.993304626 +0000 UTC m=+1200.649287100" Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.567951 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.637667 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svgvh\" (UniqueName: \"kubernetes.io/projected/002e276b-143e-4062-8530-ffaec1bdc2b8-kube-api-access-svgvh\") pod \"002e276b-143e-4062-8530-ffaec1bdc2b8\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.637831 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/002e276b-143e-4062-8530-ffaec1bdc2b8-logs\") pod \"002e276b-143e-4062-8530-ffaec1bdc2b8\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.637912 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-config-data\") pod \"002e276b-143e-4062-8530-ffaec1bdc2b8\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.641017 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-combined-ca-bundle\") pod \"002e276b-143e-4062-8530-ffaec1bdc2b8\" (UID: \"002e276b-143e-4062-8530-ffaec1bdc2b8\") " Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.641896 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/002e276b-143e-4062-8530-ffaec1bdc2b8-logs" (OuterVolumeSpecName: "logs") pod "002e276b-143e-4062-8530-ffaec1bdc2b8" (UID: "002e276b-143e-4062-8530-ffaec1bdc2b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.643228 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/002e276b-143e-4062-8530-ffaec1bdc2b8-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.665136 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/002e276b-143e-4062-8530-ffaec1bdc2b8-kube-api-access-svgvh" (OuterVolumeSpecName: "kube-api-access-svgvh") pod "002e276b-143e-4062-8530-ffaec1bdc2b8" (UID: "002e276b-143e-4062-8530-ffaec1bdc2b8"). InnerVolumeSpecName "kube-api-access-svgvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.689252 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "002e276b-143e-4062-8530-ffaec1bdc2b8" (UID: "002e276b-143e-4062-8530-ffaec1bdc2b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.689879 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-config-data" (OuterVolumeSpecName: "config-data") pod "002e276b-143e-4062-8530-ffaec1bdc2b8" (UID: "002e276b-143e-4062-8530-ffaec1bdc2b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.745558 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svgvh\" (UniqueName: \"kubernetes.io/projected/002e276b-143e-4062-8530-ffaec1bdc2b8-kube-api-access-svgvh\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.745586 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.745596 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/002e276b-143e-4062-8530-ffaec1bdc2b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.972999 4799 generic.go:334] "Generic (PLEG): container finished" podID="002e276b-143e-4062-8530-ffaec1bdc2b8" containerID="319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92" exitCode=0 Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.973041 4799 generic.go:334] "Generic (PLEG): container finished" podID="002e276b-143e-4062-8530-ffaec1bdc2b8" containerID="4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc" exitCode=143 Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.973971 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"002e276b-143e-4062-8530-ffaec1bdc2b8","Type":"ContainerDied","Data":"319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92"} Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.974002 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"002e276b-143e-4062-8530-ffaec1bdc2b8","Type":"ContainerDied","Data":"4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc"} Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.974016 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"002e276b-143e-4062-8530-ffaec1bdc2b8","Type":"ContainerDied","Data":"d4e6f97f39b33eb1f799ff6126af14bbee59cd9687555acd992d6a6472b1e48f"} Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.974036 4799 scope.go:117] "RemoveContainer" containerID="319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92" Nov 24 07:07:35 crc kubenswrapper[4799]: I1124 07:07:35.974351 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.038624 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.054967 4799 scope.go:117] "RemoveContainer" containerID="4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.061408 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.080381 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:36 crc kubenswrapper[4799]: E1124 07:07:36.081350 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="002e276b-143e-4062-8530-ffaec1bdc2b8" containerName="nova-metadata-metadata" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.081408 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="002e276b-143e-4062-8530-ffaec1bdc2b8" containerName="nova-metadata-metadata" Nov 24 07:07:36 crc kubenswrapper[4799]: E1124 07:07:36.081435 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="002e276b-143e-4062-8530-ffaec1bdc2b8" containerName="nova-metadata-log" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.081469 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="002e276b-143e-4062-8530-ffaec1bdc2b8" containerName="nova-metadata-log" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.081890 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="002e276b-143e-4062-8530-ffaec1bdc2b8" containerName="nova-metadata-log" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.081970 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="002e276b-143e-4062-8530-ffaec1bdc2b8" containerName="nova-metadata-metadata" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.083814 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.086927 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.087126 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.100289 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.107499 4799 scope.go:117] "RemoveContainer" containerID="319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92" Nov 24 07:07:36 crc kubenswrapper[4799]: E1124 07:07:36.109467 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92\": container with ID starting with 319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92 not found: ID does not exist" containerID="319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.109504 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92"} err="failed to get container status \"319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92\": rpc error: code = NotFound desc = could not find container \"319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92\": container with ID starting with 319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92 not found: ID does not exist" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.109527 4799 scope.go:117] "RemoveContainer" containerID="4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc" Nov 24 07:07:36 crc kubenswrapper[4799]: E1124 07:07:36.110015 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc\": container with ID starting with 4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc not found: ID does not exist" containerID="4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.110035 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc"} err="failed to get container status \"4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc\": rpc error: code = NotFound desc = could not find container \"4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc\": container with ID starting with 4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc not found: ID does not exist" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.110050 4799 scope.go:117] "RemoveContainer" containerID="319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.110395 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92"} err="failed to get container status \"319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92\": rpc error: code = NotFound desc = could not find container \"319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92\": container with ID starting with 319dfa805de19738d5c627de3015df3ff024041535bf2775d5ccff99e8740d92 not found: ID does not exist" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.110414 4799 scope.go:117] "RemoveContainer" containerID="4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.110727 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc"} err="failed to get container status \"4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc\": rpc error: code = NotFound desc = could not find container \"4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc\": container with ID starting with 4a209a089ed72bf9ae279f15e5e46fc1cc9203bdd7f433c974c5ddaa10e251dc not found: ID does not exist" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.155306 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-config-data\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.155405 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.155483 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.155550 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-logs\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.155584 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkzzs\" (UniqueName: \"kubernetes.io/projected/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-kube-api-access-mkzzs\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.257564 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-config-data\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.257653 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.257707 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.257763 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-logs\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.257805 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkzzs\" (UniqueName: \"kubernetes.io/projected/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-kube-api-access-mkzzs\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.259380 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-logs\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.263990 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.264009 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.264706 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-config-data\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.285269 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkzzs\" (UniqueName: \"kubernetes.io/projected/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-kube-api-access-mkzzs\") pod \"nova-metadata-0\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.413436 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.879356 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:36 crc kubenswrapper[4799]: W1124 07:07:36.889020 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d2f80ac_3117_4f0a_9a9c_7af76a286f8a.slice/crio-77d2a58d7ac6410c43fe1946b9b7c57bb250ae82236a0508e043287328167b2f WatchSource:0}: Error finding container 77d2a58d7ac6410c43fe1946b9b7c57bb250ae82236a0508e043287328167b2f: Status 404 returned error can't find the container with id 77d2a58d7ac6410c43fe1946b9b7c57bb250ae82236a0508e043287328167b2f Nov 24 07:07:36 crc kubenswrapper[4799]: I1124 07:07:36.994465 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a","Type":"ContainerStarted","Data":"77d2a58d7ac6410c43fe1946b9b7c57bb250ae82236a0508e043287328167b2f"} Nov 24 07:07:37 crc kubenswrapper[4799]: I1124 07:07:37.645250 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="002e276b-143e-4062-8530-ffaec1bdc2b8" path="/var/lib/kubelet/pods/002e276b-143e-4062-8530-ffaec1bdc2b8/volumes" Nov 24 07:07:38 crc kubenswrapper[4799]: I1124 07:07:38.010834 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a","Type":"ContainerStarted","Data":"1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43"} Nov 24 07:07:38 crc kubenswrapper[4799]: I1124 07:07:38.010918 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a","Type":"ContainerStarted","Data":"b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28"} Nov 24 07:07:38 crc kubenswrapper[4799]: I1124 07:07:38.017318 4799 generic.go:334] "Generic (PLEG): container finished" podID="650550d4-96f8-46dd-9117-d4f9c2cdcd9d" containerID="c4a15beff8279fb5630bb3a06f94fc601d98cba69dc34d057239038e1cbe6e62" exitCode=0 Nov 24 07:07:38 crc kubenswrapper[4799]: I1124 07:07:38.017381 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-k9t68" event={"ID":"650550d4-96f8-46dd-9117-d4f9c2cdcd9d","Type":"ContainerDied","Data":"c4a15beff8279fb5630bb3a06f94fc601d98cba69dc34d057239038e1cbe6e62"} Nov 24 07:07:38 crc kubenswrapper[4799]: I1124 07:07:38.038030 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.038012416 podStartE2EDuration="2.038012416s" podCreationTimestamp="2025-11-24 07:07:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:38.034485076 +0000 UTC m=+1203.690467550" watchObservedRunningTime="2025-11-24 07:07:38.038012416 +0000 UTC m=+1203.693994890" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.030773 4799 generic.go:334] "Generic (PLEG): container finished" podID="959e83f3-69f2-433b-a326-242ec1c4b226" containerID="88399e1c222bc6dd7a643d5db0cd6a4e6b60cf5019ca0bbc1ca526de8ed139d3" exitCode=0 Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.030882 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fj4rj" event={"ID":"959e83f3-69f2-433b-a326-242ec1c4b226","Type":"ContainerDied","Data":"88399e1c222bc6dd7a643d5db0cd6a4e6b60cf5019ca0bbc1ca526de8ed139d3"} Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.458191 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.501063 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.536253 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.537141 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.568805 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-tc48z"] Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.569113 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-797bbc649-tc48z" podUID="f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" containerName="dnsmasq-dns" containerID="cri-o://50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533" gracePeriod=10 Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.626994 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.628549 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-797bbc649-tc48z" podUID="f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.162:5353: connect: connection refused" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.632151 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-combined-ca-bundle\") pod \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.632237 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-scripts\") pod \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.632457 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-config-data\") pod \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.632491 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmddr\" (UniqueName: \"kubernetes.io/projected/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-kube-api-access-hmddr\") pod \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\" (UID: \"650550d4-96f8-46dd-9117-d4f9c2cdcd9d\") " Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.641353 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-scripts" (OuterVolumeSpecName: "scripts") pod "650550d4-96f8-46dd-9117-d4f9c2cdcd9d" (UID: "650550d4-96f8-46dd-9117-d4f9c2cdcd9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.646599 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-kube-api-access-hmddr" (OuterVolumeSpecName: "kube-api-access-hmddr") pod "650550d4-96f8-46dd-9117-d4f9c2cdcd9d" (UID: "650550d4-96f8-46dd-9117-d4f9c2cdcd9d"). InnerVolumeSpecName "kube-api-access-hmddr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.675383 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-config-data" (OuterVolumeSpecName: "config-data") pod "650550d4-96f8-46dd-9117-d4f9c2cdcd9d" (UID: "650550d4-96f8-46dd-9117-d4f9c2cdcd9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.682151 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "650550d4-96f8-46dd-9117-d4f9c2cdcd9d" (UID: "650550d4-96f8-46dd-9117-d4f9c2cdcd9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.735316 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.735353 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmddr\" (UniqueName: \"kubernetes.io/projected/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-kube-api-access-hmddr\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.735364 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.735372 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/650550d4-96f8-46dd-9117-d4f9c2cdcd9d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:39 crc kubenswrapper[4799]: I1124 07:07:39.753219 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.051549 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.055256 4799 generic.go:334] "Generic (PLEG): container finished" podID="f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" containerID="50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533" exitCode=0 Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.055325 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-tc48z" event={"ID":"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38","Type":"ContainerDied","Data":"50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533"} Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.055357 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-tc48z" event={"ID":"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38","Type":"ContainerDied","Data":"19d0ad5d6052c714e6dbf14a19cb13e64d58b305d7d17f1214a9067f801e7af9"} Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.055374 4799 scope.go:117] "RemoveContainer" containerID="50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.055530 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-tc48z" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.058197 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-k9t68" event={"ID":"650550d4-96f8-46dd-9117-d4f9c2cdcd9d","Type":"ContainerDied","Data":"59a7d6d0f4e4fa374070373f9908120dbcf22bb6997cd55465802dcff6d1a645"} Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.058366 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59a7d6d0f4e4fa374070373f9908120dbcf22bb6997cd55465802dcff6d1a645" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.058381 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-k9t68" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.107365 4799 scope.go:117] "RemoveContainer" containerID="589b81b002f0c3ac98becae0625aff8bf97cbfa6259f033886a8fe5e8e3f5618" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.137647 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 07:07:40 crc kubenswrapper[4799]: E1124 07:07:40.138097 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" containerName="dnsmasq-dns" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.138115 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" containerName="dnsmasq-dns" Nov 24 07:07:40 crc kubenswrapper[4799]: E1124 07:07:40.138146 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="650550d4-96f8-46dd-9117-d4f9c2cdcd9d" containerName="nova-cell1-conductor-db-sync" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.138153 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="650550d4-96f8-46dd-9117-d4f9c2cdcd9d" containerName="nova-cell1-conductor-db-sync" Nov 24 07:07:40 crc kubenswrapper[4799]: E1124 07:07:40.138162 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" containerName="init" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.138168 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" containerName="init" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.138333 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="650550d4-96f8-46dd-9117-d4f9c2cdcd9d" containerName="nova-cell1-conductor-db-sync" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.138360 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" containerName="dnsmasq-dns" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.139192 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.143833 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.151110 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-svc\") pod \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.151179 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-swift-storage-0\") pod \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.151254 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-nb\") pod \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.151318 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d2gk\" (UniqueName: \"kubernetes.io/projected/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-kube-api-access-6d2gk\") pod \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.151474 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-sb\") pod \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.151506 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-config\") pod \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\" (UID: \"f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38\") " Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.159386 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-kube-api-access-6d2gk" (OuterVolumeSpecName: "kube-api-access-6d2gk") pod "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" (UID: "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38"). InnerVolumeSpecName "kube-api-access-6d2gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.176610 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.180958 4799 scope.go:117] "RemoveContainer" containerID="50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533" Nov 24 07:07:40 crc kubenswrapper[4799]: E1124 07:07:40.182202 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533\": container with ID starting with 50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533 not found: ID does not exist" containerID="50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.182244 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533"} err="failed to get container status \"50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533\": rpc error: code = NotFound desc = could not find container \"50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533\": container with ID starting with 50bb6daaf2f46a732f7704287cc6de48f6243da92014a289effb0e42801ff533 not found: ID does not exist" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.182269 4799 scope.go:117] "RemoveContainer" containerID="589b81b002f0c3ac98becae0625aff8bf97cbfa6259f033886a8fe5e8e3f5618" Nov 24 07:07:40 crc kubenswrapper[4799]: E1124 07:07:40.185679 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"589b81b002f0c3ac98becae0625aff8bf97cbfa6259f033886a8fe5e8e3f5618\": container with ID starting with 589b81b002f0c3ac98becae0625aff8bf97cbfa6259f033886a8fe5e8e3f5618 not found: ID does not exist" containerID="589b81b002f0c3ac98becae0625aff8bf97cbfa6259f033886a8fe5e8e3f5618" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.185729 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"589b81b002f0c3ac98becae0625aff8bf97cbfa6259f033886a8fe5e8e3f5618"} err="failed to get container status \"589b81b002f0c3ac98becae0625aff8bf97cbfa6259f033886a8fe5e8e3f5618\": rpc error: code = NotFound desc = could not find container \"589b81b002f0c3ac98becae0625aff8bf97cbfa6259f033886a8fe5e8e3f5618\": container with ID starting with 589b81b002f0c3ac98becae0625aff8bf97cbfa6259f033886a8fe5e8e3f5618 not found: ID does not exist" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.197803 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.254839 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.255162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.255257 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65t49\" (UniqueName: \"kubernetes.io/projected/168550d3-6597-41ff-adf4-d20f190c7a02-kube-api-access-65t49\") pod \"nova-cell1-conductor-0\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.255440 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d2gk\" (UniqueName: \"kubernetes.io/projected/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-kube-api-access-6d2gk\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.277692 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" (UID: "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.283139 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-config" (OuterVolumeSpecName: "config") pod "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" (UID: "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.284428 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" (UID: "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.305578 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" (UID: "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.333551 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" (UID: "f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.362327 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.362565 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.362619 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65t49\" (UniqueName: \"kubernetes.io/projected/168550d3-6597-41ff-adf4-d20f190c7a02-kube-api-access-65t49\") pod \"nova-cell1-conductor-0\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.362709 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.362726 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.362738 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.362750 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.362766 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.368346 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.368675 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.381315 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65t49\" (UniqueName: \"kubernetes.io/projected/168550d3-6597-41ff-adf4-d20f190c7a02-kube-api-access-65t49\") pod \"nova-cell1-conductor-0\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.484680 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-tc48z"] Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.484937 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.518153 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-tc48z"] Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.624808 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.624827 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.640919 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.773096 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-config-data\") pod \"959e83f3-69f2-433b-a326-242ec1c4b226\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.773150 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-combined-ca-bundle\") pod \"959e83f3-69f2-433b-a326-242ec1c4b226\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.773224 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-scripts\") pod \"959e83f3-69f2-433b-a326-242ec1c4b226\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.773334 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwk9l\" (UniqueName: \"kubernetes.io/projected/959e83f3-69f2-433b-a326-242ec1c4b226-kube-api-access-dwk9l\") pod \"959e83f3-69f2-433b-a326-242ec1c4b226\" (UID: \"959e83f3-69f2-433b-a326-242ec1c4b226\") " Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.779836 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-scripts" (OuterVolumeSpecName: "scripts") pod "959e83f3-69f2-433b-a326-242ec1c4b226" (UID: "959e83f3-69f2-433b-a326-242ec1c4b226"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.780950 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/959e83f3-69f2-433b-a326-242ec1c4b226-kube-api-access-dwk9l" (OuterVolumeSpecName: "kube-api-access-dwk9l") pod "959e83f3-69f2-433b-a326-242ec1c4b226" (UID: "959e83f3-69f2-433b-a326-242ec1c4b226"). InnerVolumeSpecName "kube-api-access-dwk9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.804943 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-config-data" (OuterVolumeSpecName: "config-data") pod "959e83f3-69f2-433b-a326-242ec1c4b226" (UID: "959e83f3-69f2-433b-a326-242ec1c4b226"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.816667 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "959e83f3-69f2-433b-a326-242ec1c4b226" (UID: "959e83f3-69f2-433b-a326-242ec1c4b226"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.876615 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.877241 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.877338 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/959e83f3-69f2-433b-a326-242ec1c4b226-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.877424 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwk9l\" (UniqueName: \"kubernetes.io/projected/959e83f3-69f2-433b-a326-242ec1c4b226-kube-api-access-dwk9l\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:40 crc kubenswrapper[4799]: I1124 07:07:40.982965 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 07:07:40 crc kubenswrapper[4799]: W1124 07:07:40.984341 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod168550d3_6597_41ff_adf4_d20f190c7a02.slice/crio-b9bd0d67bff6feeb67b72850f03b07e419a11d9f67848b25fdefe0da91074088 WatchSource:0}: Error finding container b9bd0d67bff6feeb67b72850f03b07e419a11d9f67848b25fdefe0da91074088: Status 404 returned error can't find the container with id b9bd0d67bff6feeb67b72850f03b07e419a11d9f67848b25fdefe0da91074088 Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.077323 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"168550d3-6597-41ff-adf4-d20f190c7a02","Type":"ContainerStarted","Data":"b9bd0d67bff6feeb67b72850f03b07e419a11d9f67848b25fdefe0da91074088"} Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.081734 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fj4rj" event={"ID":"959e83f3-69f2-433b-a326-242ec1c4b226","Type":"ContainerDied","Data":"6a03dbaea59b59d37f24b9630ae4e68d05d1ae5f73b9763c1d53867e08b1ae63"} Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.081759 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fj4rj" Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.081772 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a03dbaea59b59d37f24b9630ae4e68d05d1ae5f73b9763c1d53867e08b1ae63" Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.209383 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.223898 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.224121 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerName="nova-api-log" containerID="cri-o://92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e" gracePeriod=30 Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.224224 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerName="nova-api-api" containerID="cri-o://f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04" gracePeriod=30 Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.233655 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.233897 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" containerName="nova-metadata-log" containerID="cri-o://b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28" gracePeriod=30 Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.234169 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" containerName="nova-metadata-metadata" containerID="cri-o://1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43" gracePeriod=30 Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.414142 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.414508 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.646749 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38" path="/var/lib/kubelet/pods/f1ef5aa0-7d62-40ae-821a-b61cb3fbbf38/volumes" Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.811721 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.901864 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-config-data\") pod \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.902072 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-logs\") pod \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.902102 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-combined-ca-bundle\") pod \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.902171 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-nova-metadata-tls-certs\") pod \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.902208 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkzzs\" (UniqueName: \"kubernetes.io/projected/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-kube-api-access-mkzzs\") pod \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\" (UID: \"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a\") " Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.902484 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-logs" (OuterVolumeSpecName: "logs") pod "6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" (UID: "6d2f80ac-3117-4f0a-9a9c-7af76a286f8a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.903061 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.926081 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-kube-api-access-mkzzs" (OuterVolumeSpecName: "kube-api-access-mkzzs") pod "6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" (UID: "6d2f80ac-3117-4f0a-9a9c-7af76a286f8a"). InnerVolumeSpecName "kube-api-access-mkzzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.954675 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" (UID: "6d2f80ac-3117-4f0a-9a9c-7af76a286f8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.960303 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-config-data" (OuterVolumeSpecName: "config-data") pod "6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" (UID: "6d2f80ac-3117-4f0a-9a9c-7af76a286f8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:41 crc kubenswrapper[4799]: I1124 07:07:41.961362 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" (UID: "6d2f80ac-3117-4f0a-9a9c-7af76a286f8a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.006705 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.006898 4799 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.006916 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkzzs\" (UniqueName: \"kubernetes.io/projected/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-kube-api-access-mkzzs\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.006927 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.092141 4799 generic.go:334] "Generic (PLEG): container finished" podID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerID="92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e" exitCode=143 Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.092427 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8","Type":"ContainerDied","Data":"92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e"} Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.093982 4799 generic.go:334] "Generic (PLEG): container finished" podID="6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" containerID="1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43" exitCode=0 Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.094006 4799 generic.go:334] "Generic (PLEG): container finished" podID="6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" containerID="b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28" exitCode=143 Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.094098 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.094095 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a","Type":"ContainerDied","Data":"1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43"} Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.094179 4799 scope.go:117] "RemoveContainer" containerID="1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.094280 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a","Type":"ContainerDied","Data":"b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28"} Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.094299 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6d2f80ac-3117-4f0a-9a9c-7af76a286f8a","Type":"ContainerDied","Data":"77d2a58d7ac6410c43fe1946b9b7c57bb250ae82236a0508e043287328167b2f"} Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.096260 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5a696130-3b34-4a3b-8ab9-677077939d0f" containerName="nova-scheduler-scheduler" containerID="cri-o://4f037689b65a1da575275f13da9f49f1ca9ea45db68ab593237cee43fc871a9b" gracePeriod=30 Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.096277 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"168550d3-6597-41ff-adf4-d20f190c7a02","Type":"ContainerStarted","Data":"192c554265f38b0137575a7bb2ad8ef843ac317c21ba65066896ec0465f061f3"} Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.096589 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.115761 4799 scope.go:117] "RemoveContainer" containerID="b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.129360 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.129347397 podStartE2EDuration="2.129347397s" podCreationTimestamp="2025-11-24 07:07:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:42.127603708 +0000 UTC m=+1207.783586182" watchObservedRunningTime="2025-11-24 07:07:42.129347397 +0000 UTC m=+1207.785329871" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.136985 4799 scope.go:117] "RemoveContainer" containerID="1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43" Nov 24 07:07:42 crc kubenswrapper[4799]: E1124 07:07:42.137385 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43\": container with ID starting with 1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43 not found: ID does not exist" containerID="1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.137428 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43"} err="failed to get container status \"1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43\": rpc error: code = NotFound desc = could not find container \"1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43\": container with ID starting with 1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43 not found: ID does not exist" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.137450 4799 scope.go:117] "RemoveContainer" containerID="b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28" Nov 24 07:07:42 crc kubenswrapper[4799]: E1124 07:07:42.137927 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28\": container with ID starting with b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28 not found: ID does not exist" containerID="b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.137954 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28"} err="failed to get container status \"b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28\": rpc error: code = NotFound desc = could not find container \"b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28\": container with ID starting with b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28 not found: ID does not exist" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.137974 4799 scope.go:117] "RemoveContainer" containerID="1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.138298 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43"} err="failed to get container status \"1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43\": rpc error: code = NotFound desc = could not find container \"1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43\": container with ID starting with 1c1092e214eca080dd02324119993d42fb0a0b73ce79f017eb6af712a2e62a43 not found: ID does not exist" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.138409 4799 scope.go:117] "RemoveContainer" containerID="b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.138834 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28"} err="failed to get container status \"b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28\": rpc error: code = NotFound desc = could not find container \"b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28\": container with ID starting with b9183423000d9d77d10e8dabf8a6485682dc8e1f9d2fd4826ef40934ee8ced28 not found: ID does not exist" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.154912 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.172482 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.198123 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:42 crc kubenswrapper[4799]: E1124 07:07:42.198726 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" containerName="nova-metadata-metadata" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.198750 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" containerName="nova-metadata-metadata" Nov 24 07:07:42 crc kubenswrapper[4799]: E1124 07:07:42.198772 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="959e83f3-69f2-433b-a326-242ec1c4b226" containerName="nova-manage" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.198781 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="959e83f3-69f2-433b-a326-242ec1c4b226" containerName="nova-manage" Nov 24 07:07:42 crc kubenswrapper[4799]: E1124 07:07:42.198805 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" containerName="nova-metadata-log" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.198813 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" containerName="nova-metadata-log" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.199091 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" containerName="nova-metadata-metadata" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.199116 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="959e83f3-69f2-433b-a326-242ec1c4b226" containerName="nova-manage" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.199145 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" containerName="nova-metadata-log" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.206036 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.206214 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.212317 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.214315 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.312198 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz6ng\" (UniqueName: \"kubernetes.io/projected/bce44781-5a13-4b7c-a9f3-022fcf589dcb-kube-api-access-lz6ng\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.312449 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.312528 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.312593 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-config-data\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.312652 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bce44781-5a13-4b7c-a9f3-022fcf589dcb-logs\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.414628 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.414700 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.414724 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-config-data\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.414753 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bce44781-5a13-4b7c-a9f3-022fcf589dcb-logs\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.414782 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz6ng\" (UniqueName: \"kubernetes.io/projected/bce44781-5a13-4b7c-a9f3-022fcf589dcb-kube-api-access-lz6ng\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.416383 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bce44781-5a13-4b7c-a9f3-022fcf589dcb-logs\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.420187 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-config-data\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.421368 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.422266 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.435340 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz6ng\" (UniqueName: \"kubernetes.io/projected/bce44781-5a13-4b7c-a9f3-022fcf589dcb-kube-api-access-lz6ng\") pod \"nova-metadata-0\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " pod="openstack/nova-metadata-0" Nov 24 07:07:42 crc kubenswrapper[4799]: I1124 07:07:42.540375 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:07:43 crc kubenswrapper[4799]: I1124 07:07:43.051939 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:07:43 crc kubenswrapper[4799]: W1124 07:07:43.058760 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbce44781_5a13_4b7c_a9f3_022fcf589dcb.slice/crio-d178da4eb487827fa15abb19b96ae96fe9f1d97c281fe32a317d6075674a7b28 WatchSource:0}: Error finding container d178da4eb487827fa15abb19b96ae96fe9f1d97c281fe32a317d6075674a7b28: Status 404 returned error can't find the container with id d178da4eb487827fa15abb19b96ae96fe9f1d97c281fe32a317d6075674a7b28 Nov 24 07:07:43 crc kubenswrapper[4799]: I1124 07:07:43.111993 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bce44781-5a13-4b7c-a9f3-022fcf589dcb","Type":"ContainerStarted","Data":"d178da4eb487827fa15abb19b96ae96fe9f1d97c281fe32a317d6075674a7b28"} Nov 24 07:07:43 crc kubenswrapper[4799]: I1124 07:07:43.647533 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d2f80ac-3117-4f0a-9a9c-7af76a286f8a" path="/var/lib/kubelet/pods/6d2f80ac-3117-4f0a-9a9c-7af76a286f8a/volumes" Nov 24 07:07:44 crc kubenswrapper[4799]: I1124 07:07:44.125661 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bce44781-5a13-4b7c-a9f3-022fcf589dcb","Type":"ContainerStarted","Data":"9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34"} Nov 24 07:07:44 crc kubenswrapper[4799]: I1124 07:07:44.127697 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bce44781-5a13-4b7c-a9f3-022fcf589dcb","Type":"ContainerStarted","Data":"2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9"} Nov 24 07:07:44 crc kubenswrapper[4799]: I1124 07:07:44.158099 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.158074816 podStartE2EDuration="2.158074816s" podCreationTimestamp="2025-11-24 07:07:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:44.153999591 +0000 UTC m=+1209.809982105" watchObservedRunningTime="2025-11-24 07:07:44.158074816 +0000 UTC m=+1209.814057310" Nov 24 07:07:44 crc kubenswrapper[4799]: E1124 07:07:44.630100 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f037689b65a1da575275f13da9f49f1ca9ea45db68ab593237cee43fc871a9b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:07:44 crc kubenswrapper[4799]: E1124 07:07:44.632368 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f037689b65a1da575275f13da9f49f1ca9ea45db68ab593237cee43fc871a9b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:07:44 crc kubenswrapper[4799]: E1124 07:07:44.634290 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f037689b65a1da575275f13da9f49f1ca9ea45db68ab593237cee43fc871a9b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:07:44 crc kubenswrapper[4799]: E1124 07:07:44.634448 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="5a696130-3b34-4a3b-8ab9-677077939d0f" containerName="nova-scheduler-scheduler" Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.167315 4799 generic.go:334] "Generic (PLEG): container finished" podID="5a696130-3b34-4a3b-8ab9-677077939d0f" containerID="4f037689b65a1da575275f13da9f49f1ca9ea45db68ab593237cee43fc871a9b" exitCode=0 Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.167700 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a696130-3b34-4a3b-8ab9-677077939d0f","Type":"ContainerDied","Data":"4f037689b65a1da575275f13da9f49f1ca9ea45db68ab593237cee43fc871a9b"} Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.167734 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a696130-3b34-4a3b-8ab9-677077939d0f","Type":"ContainerDied","Data":"cf3af3c3025f3af56dfb33fffc7204978605c8c7d5a0d68f6749e9d4b731993f"} Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.167746 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf3af3c3025f3af56dfb33fffc7204978605c8c7d5a0d68f6749e9d4b731993f" Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.177487 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.296861 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-combined-ca-bundle\") pod \"5a696130-3b34-4a3b-8ab9-677077939d0f\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.297233 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-config-data\") pod \"5a696130-3b34-4a3b-8ab9-677077939d0f\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.297476 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkz8h\" (UniqueName: \"kubernetes.io/projected/5a696130-3b34-4a3b-8ab9-677077939d0f-kube-api-access-xkz8h\") pod \"5a696130-3b34-4a3b-8ab9-677077939d0f\" (UID: \"5a696130-3b34-4a3b-8ab9-677077939d0f\") " Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.302988 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a696130-3b34-4a3b-8ab9-677077939d0f-kube-api-access-xkz8h" (OuterVolumeSpecName: "kube-api-access-xkz8h") pod "5a696130-3b34-4a3b-8ab9-677077939d0f" (UID: "5a696130-3b34-4a3b-8ab9-677077939d0f"). InnerVolumeSpecName "kube-api-access-xkz8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.334154 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-config-data" (OuterVolumeSpecName: "config-data") pod "5a696130-3b34-4a3b-8ab9-677077939d0f" (UID: "5a696130-3b34-4a3b-8ab9-677077939d0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.334499 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a696130-3b34-4a3b-8ab9-677077939d0f" (UID: "5a696130-3b34-4a3b-8ab9-677077939d0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.399764 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.399801 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a696130-3b34-4a3b-8ab9-677077939d0f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:46 crc kubenswrapper[4799]: I1124 07:07:46.399810 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkz8h\" (UniqueName: \"kubernetes.io/projected/5a696130-3b34-4a3b-8ab9-677077939d0f-kube-api-access-xkz8h\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.049882 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.181914 4799 generic.go:334] "Generic (PLEG): container finished" podID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerID="f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04" exitCode=0 Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.182038 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.182045 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.182025 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8","Type":"ContainerDied","Data":"f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04"} Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.182179 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8","Type":"ContainerDied","Data":"af5f7cd771c629cbcd61bb98a6da710202ea0378475baacf83fb3a7d21b49810"} Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.182208 4799 scope.go:117] "RemoveContainer" containerID="f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.212610 4799 scope.go:117] "RemoveContainer" containerID="92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.217524 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx792\" (UniqueName: \"kubernetes.io/projected/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-kube-api-access-sx792\") pod \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.217592 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-config-data\") pod \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.217711 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-logs\") pod \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.217840 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-combined-ca-bundle\") pod \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\" (UID: \"e1f2ac4d-9072-4dab-aa13-16a0dc688bb8\") " Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.219193 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-logs" (OuterVolumeSpecName: "logs") pod "e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" (UID: "e1f2ac4d-9072-4dab-aa13-16a0dc688bb8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.225388 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.238203 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-kube-api-access-sx792" (OuterVolumeSpecName: "kube-api-access-sx792") pod "e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" (UID: "e1f2ac4d-9072-4dab-aa13-16a0dc688bb8"). InnerVolumeSpecName "kube-api-access-sx792". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.253507 4799 scope.go:117] "RemoveContainer" containerID="f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04" Nov 24 07:07:47 crc kubenswrapper[4799]: E1124 07:07:47.254112 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04\": container with ID starting with f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04 not found: ID does not exist" containerID="f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.254182 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04"} err="failed to get container status \"f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04\": rpc error: code = NotFound desc = could not find container \"f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04\": container with ID starting with f8f4a52004c2497f11f2644ee71ee5bc8b262434755231012cc42175f73b3f04 not found: ID does not exist" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.254221 4799 scope.go:117] "RemoveContainer" containerID="92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e" Nov 24 07:07:47 crc kubenswrapper[4799]: E1124 07:07:47.254545 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e\": container with ID starting with 92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e not found: ID does not exist" containerID="92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.254584 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e"} err="failed to get container status \"92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e\": rpc error: code = NotFound desc = could not find container \"92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e\": container with ID starting with 92f426a0af465c2a2aa6ef63a12389802a0a904c48da4f1a9b81a65bacd5772e not found: ID does not exist" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.257089 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.260768 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" (UID: "e1f2ac4d-9072-4dab-aa13-16a0dc688bb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.279502 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-config-data" (OuterVolumeSpecName: "config-data") pod "e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" (UID: "e1f2ac4d-9072-4dab-aa13-16a0dc688bb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.291450 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:07:47 crc kubenswrapper[4799]: E1124 07:07:47.292157 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a696130-3b34-4a3b-8ab9-677077939d0f" containerName="nova-scheduler-scheduler" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.292242 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a696130-3b34-4a3b-8ab9-677077939d0f" containerName="nova-scheduler-scheduler" Nov 24 07:07:47 crc kubenswrapper[4799]: E1124 07:07:47.292326 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerName="nova-api-api" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.292379 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerName="nova-api-api" Nov 24 07:07:47 crc kubenswrapper[4799]: E1124 07:07:47.292442 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerName="nova-api-log" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.292510 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerName="nova-api-log" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.292891 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerName="nova-api-log" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.292975 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" containerName="nova-api-api" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.293044 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a696130-3b34-4a3b-8ab9-677077939d0f" containerName="nova-scheduler-scheduler" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.293878 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.296328 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.298588 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.320092 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.320135 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx792\" (UniqueName: \"kubernetes.io/projected/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-kube-api-access-sx792\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.320147 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.320162 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.422367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lhqp\" (UniqueName: \"kubernetes.io/projected/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-kube-api-access-6lhqp\") pod \"nova-scheduler-0\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.422479 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-config-data\") pod \"nova-scheduler-0\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.422507 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.523954 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lhqp\" (UniqueName: \"kubernetes.io/projected/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-kube-api-access-6lhqp\") pod \"nova-scheduler-0\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.524008 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-config-data\") pod \"nova-scheduler-0\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.524038 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.531998 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.534899 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.535327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-config-data\") pod \"nova-scheduler-0\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.540713 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.540788 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.541382 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.549747 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lhqp\" (UniqueName: \"kubernetes.io/projected/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-kube-api-access-6lhqp\") pod \"nova-scheduler-0\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.564019 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.566232 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.569172 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.580110 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.643405 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a696130-3b34-4a3b-8ab9-677077939d0f" path="/var/lib/kubelet/pods/5a696130-3b34-4a3b-8ab9-677077939d0f/volumes" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.644617 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1f2ac4d-9072-4dab-aa13-16a0dc688bb8" path="/var/lib/kubelet/pods/e1f2ac4d-9072-4dab-aa13-16a0dc688bb8/volumes" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.644964 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.727572 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-config-data\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.728276 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbed3e4-9342-4cd7-a142-9926ad732782-logs\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.728315 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.728408 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqkls\" (UniqueName: \"kubernetes.io/projected/8cbed3e4-9342-4cd7-a142-9926ad732782-kube-api-access-mqkls\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.830439 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbed3e4-9342-4cd7-a142-9926ad732782-logs\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.830490 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.830543 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqkls\" (UniqueName: \"kubernetes.io/projected/8cbed3e4-9342-4cd7-a142-9926ad732782-kube-api-access-mqkls\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.830574 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-config-data\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.830975 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbed3e4-9342-4cd7-a142-9926ad732782-logs\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.835765 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.849576 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-config-data\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.850150 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqkls\" (UniqueName: \"kubernetes.io/projected/8cbed3e4-9342-4cd7-a142-9926ad732782-kube-api-access-mqkls\") pod \"nova-api-0\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " pod="openstack/nova-api-0" Nov 24 07:07:47 crc kubenswrapper[4799]: I1124 07:07:47.886005 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:07:48 crc kubenswrapper[4799]: I1124 07:07:48.118680 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:07:48 crc kubenswrapper[4799]: I1124 07:07:48.197130 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58bc8e5c-1b62-4f9e-8e0e-83e40a318193","Type":"ContainerStarted","Data":"59d2ba4f8b07df33597c9fc41405b43cda395c0d41e140e07c40d1735d7d587e"} Nov 24 07:07:48 crc kubenswrapper[4799]: I1124 07:07:48.350593 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:07:48 crc kubenswrapper[4799]: W1124 07:07:48.351319 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cbed3e4_9342_4cd7_a142_9926ad732782.slice/crio-5ba62f83eefe7036ffa500cc074774500f001e31e2784b12d9d7358f4a82e1f3 WatchSource:0}: Error finding container 5ba62f83eefe7036ffa500cc074774500f001e31e2784b12d9d7358f4a82e1f3: Status 404 returned error can't find the container with id 5ba62f83eefe7036ffa500cc074774500f001e31e2784b12d9d7358f4a82e1f3 Nov 24 07:07:49 crc kubenswrapper[4799]: I1124 07:07:49.212490 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58bc8e5c-1b62-4f9e-8e0e-83e40a318193","Type":"ContainerStarted","Data":"81617cd15db8bcf8cda5a02156d69677d41b237ae096f7be2333e6cb47100016"} Nov 24 07:07:49 crc kubenswrapper[4799]: I1124 07:07:49.217006 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cbed3e4-9342-4cd7-a142-9926ad732782","Type":"ContainerStarted","Data":"24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93"} Nov 24 07:07:49 crc kubenswrapper[4799]: I1124 07:07:49.217074 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cbed3e4-9342-4cd7-a142-9926ad732782","Type":"ContainerStarted","Data":"8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0"} Nov 24 07:07:49 crc kubenswrapper[4799]: I1124 07:07:49.217106 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cbed3e4-9342-4cd7-a142-9926ad732782","Type":"ContainerStarted","Data":"5ba62f83eefe7036ffa500cc074774500f001e31e2784b12d9d7358f4a82e1f3"} Nov 24 07:07:49 crc kubenswrapper[4799]: I1124 07:07:49.240274 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.240246925 podStartE2EDuration="2.240246925s" podCreationTimestamp="2025-11-24 07:07:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:49.233051971 +0000 UTC m=+1214.889034445" watchObservedRunningTime="2025-11-24 07:07:49.240246925 +0000 UTC m=+1214.896229439" Nov 24 07:07:49 crc kubenswrapper[4799]: I1124 07:07:49.263579 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.263556697 podStartE2EDuration="2.263556697s" podCreationTimestamp="2025-11-24 07:07:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:07:49.254999764 +0000 UTC m=+1214.910982278" watchObservedRunningTime="2025-11-24 07:07:49.263556697 +0000 UTC m=+1214.919539171" Nov 24 07:07:50 crc kubenswrapper[4799]: I1124 07:07:50.321950 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 07:07:50 crc kubenswrapper[4799]: I1124 07:07:50.525063 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 07:07:52 crc kubenswrapper[4799]: I1124 07:07:52.541317 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 07:07:52 crc kubenswrapper[4799]: I1124 07:07:52.541872 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 07:07:52 crc kubenswrapper[4799]: I1124 07:07:52.646143 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 07:07:53 crc kubenswrapper[4799]: I1124 07:07:53.558139 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:07:53 crc kubenswrapper[4799]: I1124 07:07:53.558458 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:07:54 crc kubenswrapper[4799]: I1124 07:07:54.614505 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:07:54 crc kubenswrapper[4799]: I1124 07:07:54.614964 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="04342f4b-48ae-4c45-ad6b-947b30b9de75" containerName="kube-state-metrics" containerID="cri-o://7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c" gracePeriod=30 Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.140297 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.141481 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbwwq\" (UniqueName: \"kubernetes.io/projected/04342f4b-48ae-4c45-ad6b-947b30b9de75-kube-api-access-bbwwq\") pod \"04342f4b-48ae-4c45-ad6b-947b30b9de75\" (UID: \"04342f4b-48ae-4c45-ad6b-947b30b9de75\") " Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.149722 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04342f4b-48ae-4c45-ad6b-947b30b9de75-kube-api-access-bbwwq" (OuterVolumeSpecName: "kube-api-access-bbwwq") pod "04342f4b-48ae-4c45-ad6b-947b30b9de75" (UID: "04342f4b-48ae-4c45-ad6b-947b30b9de75"). InnerVolumeSpecName "kube-api-access-bbwwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.245051 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbwwq\" (UniqueName: \"kubernetes.io/projected/04342f4b-48ae-4c45-ad6b-947b30b9de75-kube-api-access-bbwwq\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.280710 4799 generic.go:334] "Generic (PLEG): container finished" podID="04342f4b-48ae-4c45-ad6b-947b30b9de75" containerID="7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c" exitCode=2 Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.280812 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"04342f4b-48ae-4c45-ad6b-947b30b9de75","Type":"ContainerDied","Data":"7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c"} Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.280879 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"04342f4b-48ae-4c45-ad6b-947b30b9de75","Type":"ContainerDied","Data":"80ffe4ba2b05d66cbac89e0520e04e2880c6f22f2ff8744e1caeb369043cff2f"} Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.280889 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.280912 4799 scope.go:117] "RemoveContainer" containerID="7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.305837 4799 scope.go:117] "RemoveContainer" containerID="7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c" Nov 24 07:07:55 crc kubenswrapper[4799]: E1124 07:07:55.306579 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c\": container with ID starting with 7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c not found: ID does not exist" containerID="7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.306641 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c"} err="failed to get container status \"7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c\": rpc error: code = NotFound desc = could not find container \"7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c\": container with ID starting with 7b4b0164dbb4eae5b3d743c57c2a436bfe1f57d7cde5d599dcdbd3b4739b957c not found: ID does not exist" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.339967 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.374502 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.380526 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:07:55 crc kubenswrapper[4799]: E1124 07:07:55.381159 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04342f4b-48ae-4c45-ad6b-947b30b9de75" containerName="kube-state-metrics" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.381186 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="04342f4b-48ae-4c45-ad6b-947b30b9de75" containerName="kube-state-metrics" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.381436 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="04342f4b-48ae-4c45-ad6b-947b30b9de75" containerName="kube-state-metrics" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.382504 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.385246 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.385519 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.388036 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.560511 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.560617 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.560749 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.560800 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68hmq\" (UniqueName: \"kubernetes.io/projected/3cd44457-1418-4788-a029-82c54c917997-kube-api-access-68hmq\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.640426 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04342f4b-48ae-4c45-ad6b-947b30b9de75" path="/var/lib/kubelet/pods/04342f4b-48ae-4c45-ad6b-947b30b9de75/volumes" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.663123 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.663256 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.663401 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.663474 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68hmq\" (UniqueName: \"kubernetes.io/projected/3cd44457-1418-4788-a029-82c54c917997-kube-api-access-68hmq\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.669202 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.669576 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.669977 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.680453 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68hmq\" (UniqueName: \"kubernetes.io/projected/3cd44457-1418-4788-a029-82c54c917997-kube-api-access-68hmq\") pod \"kube-state-metrics-0\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " pod="openstack/kube-state-metrics-0" Nov 24 07:07:55 crc kubenswrapper[4799]: I1124 07:07:55.700744 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:07:56 crc kubenswrapper[4799]: I1124 07:07:56.230025 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:07:56 crc kubenswrapper[4799]: I1124 07:07:56.296456 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3cd44457-1418-4788-a029-82c54c917997","Type":"ContainerStarted","Data":"63de444daa10203a26673c85620a0d616f43b8136c3953d59df36db67362f926"} Nov 24 07:07:56 crc kubenswrapper[4799]: I1124 07:07:56.396912 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:07:56 crc kubenswrapper[4799]: I1124 07:07:56.397726 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="ceilometer-central-agent" containerID="cri-o://f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663" gracePeriod=30 Nov 24 07:07:56 crc kubenswrapper[4799]: I1124 07:07:56.397811 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="sg-core" containerID="cri-o://ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786" gracePeriod=30 Nov 24 07:07:56 crc kubenswrapper[4799]: I1124 07:07:56.397814 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="proxy-httpd" containerID="cri-o://076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6" gracePeriod=30 Nov 24 07:07:56 crc kubenswrapper[4799]: I1124 07:07:56.397809 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="ceilometer-notification-agent" containerID="cri-o://f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31" gracePeriod=30 Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.308837 4799 generic.go:334] "Generic (PLEG): container finished" podID="2119f5af-764e-4301-9b73-e44b3771c62a" containerID="076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6" exitCode=0 Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.309241 4799 generic.go:334] "Generic (PLEG): container finished" podID="2119f5af-764e-4301-9b73-e44b3771c62a" containerID="ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786" exitCode=2 Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.309250 4799 generic.go:334] "Generic (PLEG): container finished" podID="2119f5af-764e-4301-9b73-e44b3771c62a" containerID="f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663" exitCode=0 Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.308931 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2119f5af-764e-4301-9b73-e44b3771c62a","Type":"ContainerDied","Data":"076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6"} Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.309332 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2119f5af-764e-4301-9b73-e44b3771c62a","Type":"ContainerDied","Data":"ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786"} Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.309352 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2119f5af-764e-4301-9b73-e44b3771c62a","Type":"ContainerDied","Data":"f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663"} Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.312341 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3cd44457-1418-4788-a029-82c54c917997","Type":"ContainerStarted","Data":"e7c302a760dcca397dd206e8bf79d8516ca32ba49a1ef37a43e67e1893d9366d"} Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.313050 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.334221 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.9362265650000001 podStartE2EDuration="2.334202719s" podCreationTimestamp="2025-11-24 07:07:55 +0000 UTC" firstStartedPulling="2025-11-24 07:07:56.241835721 +0000 UTC m=+1221.897818235" lastFinishedPulling="2025-11-24 07:07:56.639811915 +0000 UTC m=+1222.295794389" observedRunningTime="2025-11-24 07:07:57.332658695 +0000 UTC m=+1222.988641169" watchObservedRunningTime="2025-11-24 07:07:57.334202719 +0000 UTC m=+1222.990185193" Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.650192 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.684249 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.886651 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:07:57 crc kubenswrapper[4799]: I1124 07:07:57.888250 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:07:58 crc kubenswrapper[4799]: I1124 07:07:58.350584 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 07:07:58 crc kubenswrapper[4799]: I1124 07:07:58.969127 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 07:07:58 crc kubenswrapper[4799]: I1124 07:07:58.969184 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 07:07:58 crc kubenswrapper[4799]: I1124 07:07:58.986140 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.153042 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-combined-ca-bundle\") pod \"2119f5af-764e-4301-9b73-e44b3771c62a\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.153207 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-sg-core-conf-yaml\") pod \"2119f5af-764e-4301-9b73-e44b3771c62a\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.153292 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-scripts\") pod \"2119f5af-764e-4301-9b73-e44b3771c62a\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.153464 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-run-httpd\") pod \"2119f5af-764e-4301-9b73-e44b3771c62a\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.153596 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv7j2\" (UniqueName: \"kubernetes.io/projected/2119f5af-764e-4301-9b73-e44b3771c62a-kube-api-access-tv7j2\") pod \"2119f5af-764e-4301-9b73-e44b3771c62a\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.153666 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-config-data\") pod \"2119f5af-764e-4301-9b73-e44b3771c62a\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.153718 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-log-httpd\") pod \"2119f5af-764e-4301-9b73-e44b3771c62a\" (UID: \"2119f5af-764e-4301-9b73-e44b3771c62a\") " Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.154533 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2119f5af-764e-4301-9b73-e44b3771c62a" (UID: "2119f5af-764e-4301-9b73-e44b3771c62a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.155075 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2119f5af-764e-4301-9b73-e44b3771c62a" (UID: "2119f5af-764e-4301-9b73-e44b3771c62a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.164156 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2119f5af-764e-4301-9b73-e44b3771c62a-kube-api-access-tv7j2" (OuterVolumeSpecName: "kube-api-access-tv7j2") pod "2119f5af-764e-4301-9b73-e44b3771c62a" (UID: "2119f5af-764e-4301-9b73-e44b3771c62a"). InnerVolumeSpecName "kube-api-access-tv7j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.165220 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-scripts" (OuterVolumeSpecName: "scripts") pod "2119f5af-764e-4301-9b73-e44b3771c62a" (UID: "2119f5af-764e-4301-9b73-e44b3771c62a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.196919 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2119f5af-764e-4301-9b73-e44b3771c62a" (UID: "2119f5af-764e-4301-9b73-e44b3771c62a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.256666 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.256706 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.256716 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.256726 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv7j2\" (UniqueName: \"kubernetes.io/projected/2119f5af-764e-4301-9b73-e44b3771c62a-kube-api-access-tv7j2\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.256736 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2119f5af-764e-4301-9b73-e44b3771c62a-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.257189 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2119f5af-764e-4301-9b73-e44b3771c62a" (UID: "2119f5af-764e-4301-9b73-e44b3771c62a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.291747 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-config-data" (OuterVolumeSpecName: "config-data") pod "2119f5af-764e-4301-9b73-e44b3771c62a" (UID: "2119f5af-764e-4301-9b73-e44b3771c62a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.331516 4799 generic.go:334] "Generic (PLEG): container finished" podID="2119f5af-764e-4301-9b73-e44b3771c62a" containerID="f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31" exitCode=0 Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.331552 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2119f5af-764e-4301-9b73-e44b3771c62a","Type":"ContainerDied","Data":"f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31"} Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.331592 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2119f5af-764e-4301-9b73-e44b3771c62a","Type":"ContainerDied","Data":"81a47571abcba2e0656a1ce37b514272f6eb33b0b6fba845d04cf613247158f1"} Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.331597 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.331613 4799 scope.go:117] "RemoveContainer" containerID="076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.358674 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.359600 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2119f5af-764e-4301-9b73-e44b3771c62a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.362931 4799 scope.go:117] "RemoveContainer" containerID="ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.371365 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.384317 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.396801 4799 scope.go:117] "RemoveContainer" containerID="f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.411369 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:07:59 crc kubenswrapper[4799]: E1124 07:07:59.411829 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="proxy-httpd" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.411858 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="proxy-httpd" Nov 24 07:07:59 crc kubenswrapper[4799]: E1124 07:07:59.411881 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="sg-core" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.411888 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="sg-core" Nov 24 07:07:59 crc kubenswrapper[4799]: E1124 07:07:59.411918 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="ceilometer-central-agent" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.411926 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="ceilometer-central-agent" Nov 24 07:07:59 crc kubenswrapper[4799]: E1124 07:07:59.411939 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="ceilometer-notification-agent" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.411945 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="ceilometer-notification-agent" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.412169 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="sg-core" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.412186 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="ceilometer-central-agent" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.412196 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="ceilometer-notification-agent" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.412211 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" containerName="proxy-httpd" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.414014 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.418331 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.418470 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.418591 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.426673 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.428986 4799 scope.go:117] "RemoveContainer" containerID="f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.464990 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.465120 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.465961 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.466048 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-scripts\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.466075 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-config-data\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.466134 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkqkk\" (UniqueName: \"kubernetes.io/projected/133ef3af-ba10-4ad0-bee3-dce60aac408f-kube-api-access-hkqkk\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.466296 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-run-httpd\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.466346 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-log-httpd\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.470336 4799 scope.go:117] "RemoveContainer" containerID="076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6" Nov 24 07:07:59 crc kubenswrapper[4799]: E1124 07:07:59.473277 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6\": container with ID starting with 076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6 not found: ID does not exist" containerID="076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.473393 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6"} err="failed to get container status \"076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6\": rpc error: code = NotFound desc = could not find container \"076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6\": container with ID starting with 076801a5bc28659befe36091b111ab02986ca2908c51297f91a794727e1e31f6 not found: ID does not exist" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.473427 4799 scope.go:117] "RemoveContainer" containerID="ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786" Nov 24 07:07:59 crc kubenswrapper[4799]: E1124 07:07:59.473767 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786\": container with ID starting with ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786 not found: ID does not exist" containerID="ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.473796 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786"} err="failed to get container status \"ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786\": rpc error: code = NotFound desc = could not find container \"ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786\": container with ID starting with ee5325a03e97728fdb3a6cd19d3cd6d709e7f3fcd9e9379ef0a3bddd9f08d786 not found: ID does not exist" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.473811 4799 scope.go:117] "RemoveContainer" containerID="f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31" Nov 24 07:07:59 crc kubenswrapper[4799]: E1124 07:07:59.474133 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31\": container with ID starting with f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31 not found: ID does not exist" containerID="f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.474192 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31"} err="failed to get container status \"f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31\": rpc error: code = NotFound desc = could not find container \"f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31\": container with ID starting with f66d7e8de61013dc11186b76dc87653c285894afdaa3da695df48f4ff751de31 not found: ID does not exist" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.474227 4799 scope.go:117] "RemoveContainer" containerID="f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663" Nov 24 07:07:59 crc kubenswrapper[4799]: E1124 07:07:59.474528 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663\": container with ID starting with f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663 not found: ID does not exist" containerID="f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.474553 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663"} err="failed to get container status \"f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663\": rpc error: code = NotFound desc = could not find container \"f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663\": container with ID starting with f24924af98689f5258bec338a9eb3c581030e57b436f675292de893e7b06a663 not found: ID does not exist" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.567914 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-run-httpd\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.568291 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-log-httpd\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.568408 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.568446 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.568480 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.568520 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-scripts\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.568541 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-config-data\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.568575 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkqkk\" (UniqueName: \"kubernetes.io/projected/133ef3af-ba10-4ad0-bee3-dce60aac408f-kube-api-access-hkqkk\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.568782 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-run-httpd\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.572990 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-log-httpd\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.575458 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.575805 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-scripts\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.578240 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.578588 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.579672 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-config-data\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.588362 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkqkk\" (UniqueName: \"kubernetes.io/projected/133ef3af-ba10-4ad0-bee3-dce60aac408f-kube-api-access-hkqkk\") pod \"ceilometer-0\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " pod="openstack/ceilometer-0" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.650947 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2119f5af-764e-4301-9b73-e44b3771c62a" path="/var/lib/kubelet/pods/2119f5af-764e-4301-9b73-e44b3771c62a/volumes" Nov 24 07:07:59 crc kubenswrapper[4799]: I1124 07:07:59.749928 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:08:00 crc kubenswrapper[4799]: I1124 07:08:00.336606 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:00 crc kubenswrapper[4799]: W1124 07:08:00.351161 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod133ef3af_ba10_4ad0_bee3_dce60aac408f.slice/crio-115bbe449ce97aa5b7ff52dcaea8564db9cacc851c1305ad9cfec244201a7579 WatchSource:0}: Error finding container 115bbe449ce97aa5b7ff52dcaea8564db9cacc851c1305ad9cfec244201a7579: Status 404 returned error can't find the container with id 115bbe449ce97aa5b7ff52dcaea8564db9cacc851c1305ad9cfec244201a7579 Nov 24 07:08:01 crc kubenswrapper[4799]: I1124 07:08:01.381390 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"133ef3af-ba10-4ad0-bee3-dce60aac408f","Type":"ContainerStarted","Data":"8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd"} Nov 24 07:08:01 crc kubenswrapper[4799]: I1124 07:08:01.382196 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"133ef3af-ba10-4ad0-bee3-dce60aac408f","Type":"ContainerStarted","Data":"115bbe449ce97aa5b7ff52dcaea8564db9cacc851c1305ad9cfec244201a7579"} Nov 24 07:08:02 crc kubenswrapper[4799]: I1124 07:08:02.391315 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"133ef3af-ba10-4ad0-bee3-dce60aac408f","Type":"ContainerStarted","Data":"7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1"} Nov 24 07:08:02 crc kubenswrapper[4799]: I1124 07:08:02.545760 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 07:08:02 crc kubenswrapper[4799]: I1124 07:08:02.546764 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 07:08:02 crc kubenswrapper[4799]: I1124 07:08:02.551119 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 07:08:03 crc kubenswrapper[4799]: I1124 07:08:03.412803 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"133ef3af-ba10-4ad0-bee3-dce60aac408f","Type":"ContainerStarted","Data":"f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb"} Nov 24 07:08:03 crc kubenswrapper[4799]: I1124 07:08:03.418721 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.358679 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.421032 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"133ef3af-ba10-4ad0-bee3-dce60aac408f","Type":"ContainerStarted","Data":"a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d"} Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.424109 4799 generic.go:334] "Generic (PLEG): container finished" podID="4c989b53-8318-4b5d-b893-20c0d1e98cd4" containerID="28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b" exitCode=137 Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.424168 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.424260 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4c989b53-8318-4b5d-b893-20c0d1e98cd4","Type":"ContainerDied","Data":"28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b"} Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.424331 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4c989b53-8318-4b5d-b893-20c0d1e98cd4","Type":"ContainerDied","Data":"145996eed3db13e8d38ac0c13cf4336848c1f7bb23b558439ebbaa32800e3113"} Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.424354 4799 scope.go:117] "RemoveContainer" containerID="28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.445103 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.767177008 podStartE2EDuration="5.445075076s" podCreationTimestamp="2025-11-24 07:07:59 +0000 UTC" firstStartedPulling="2025-11-24 07:08:00.35528092 +0000 UTC m=+1226.011263404" lastFinishedPulling="2025-11-24 07:08:04.033178988 +0000 UTC m=+1229.689161472" observedRunningTime="2025-11-24 07:08:04.444079878 +0000 UTC m=+1230.100062352" watchObservedRunningTime="2025-11-24 07:08:04.445075076 +0000 UTC m=+1230.101057560" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.462409 4799 scope.go:117] "RemoveContainer" containerID="28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b" Nov 24 07:08:04 crc kubenswrapper[4799]: E1124 07:08:04.463279 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b\": container with ID starting with 28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b not found: ID does not exist" containerID="28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.463330 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b"} err="failed to get container status \"28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b\": rpc error: code = NotFound desc = could not find container \"28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b\": container with ID starting with 28adfb74a1eef335c2fc4b742ba6745e94ed0fd943f656348474ee73eba18d3b not found: ID does not exist" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.481784 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt2sf\" (UniqueName: \"kubernetes.io/projected/4c989b53-8318-4b5d-b893-20c0d1e98cd4-kube-api-access-wt2sf\") pod \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.483034 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-combined-ca-bundle\") pod \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.483072 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-config-data\") pod \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\" (UID: \"4c989b53-8318-4b5d-b893-20c0d1e98cd4\") " Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.487486 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c989b53-8318-4b5d-b893-20c0d1e98cd4-kube-api-access-wt2sf" (OuterVolumeSpecName: "kube-api-access-wt2sf") pod "4c989b53-8318-4b5d-b893-20c0d1e98cd4" (UID: "4c989b53-8318-4b5d-b893-20c0d1e98cd4"). InnerVolumeSpecName "kube-api-access-wt2sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.512949 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c989b53-8318-4b5d-b893-20c0d1e98cd4" (UID: "4c989b53-8318-4b5d-b893-20c0d1e98cd4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.519381 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-config-data" (OuterVolumeSpecName: "config-data") pod "4c989b53-8318-4b5d-b893-20c0d1e98cd4" (UID: "4c989b53-8318-4b5d-b893-20c0d1e98cd4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.588711 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt2sf\" (UniqueName: \"kubernetes.io/projected/4c989b53-8318-4b5d-b893-20c0d1e98cd4-kube-api-access-wt2sf\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.589873 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.589897 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c989b53-8318-4b5d-b893-20c0d1e98cd4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.767612 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.789555 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.809519 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:08:04 crc kubenswrapper[4799]: E1124 07:08:04.810251 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c989b53-8318-4b5d-b893-20c0d1e98cd4" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.810284 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c989b53-8318-4b5d-b893-20c0d1e98cd4" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.810665 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c989b53-8318-4b5d-b893-20c0d1e98cd4" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.813028 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.816793 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.817321 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.817972 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.835329 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.998866 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.999059 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.999136 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.999179 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8fx9\" (UniqueName: \"kubernetes.io/projected/3feaa8e4-537c-44d2-ae92-ba999eb1626c-kube-api-access-n8fx9\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:04 crc kubenswrapper[4799]: I1124 07:08:04.999215 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.117140 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.117331 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.117405 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.117443 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8fx9\" (UniqueName: \"kubernetes.io/projected/3feaa8e4-537c-44d2-ae92-ba999eb1626c-kube-api-access-n8fx9\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.117475 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.124126 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.124467 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.125272 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.128126 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.136399 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8fx9\" (UniqueName: \"kubernetes.io/projected/3feaa8e4-537c-44d2-ae92-ba999eb1626c-kube-api-access-n8fx9\") pod \"nova-cell1-novncproxy-0\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.435768 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.437913 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.644912 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c989b53-8318-4b5d-b893-20c0d1e98cd4" path="/var/lib/kubelet/pods/4c989b53-8318-4b5d-b893-20c0d1e98cd4/volumes" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.718383 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 07:08:05 crc kubenswrapper[4799]: I1124 07:08:05.962195 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:08:06 crc kubenswrapper[4799]: I1124 07:08:06.450446 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3feaa8e4-537c-44d2-ae92-ba999eb1626c","Type":"ContainerStarted","Data":"330d7990c1e4ba99f0cea6c8c2d506420220e3481e87800ebda2781888dd174a"} Nov 24 07:08:06 crc kubenswrapper[4799]: I1124 07:08:06.451080 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3feaa8e4-537c-44d2-ae92-ba999eb1626c","Type":"ContainerStarted","Data":"57719c24ba2aa35603649c3f2a0655efe9f343e4b6858acf1844fc6dd510aac8"} Nov 24 07:08:06 crc kubenswrapper[4799]: I1124 07:08:06.472483 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.472452568 podStartE2EDuration="2.472452568s" podCreationTimestamp="2025-11-24 07:08:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:08:06.466388436 +0000 UTC m=+1232.122370910" watchObservedRunningTime="2025-11-24 07:08:06.472452568 +0000 UTC m=+1232.128435052" Nov 24 07:08:07 crc kubenswrapper[4799]: I1124 07:08:07.889814 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 07:08:07 crc kubenswrapper[4799]: I1124 07:08:07.891312 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 07:08:07 crc kubenswrapper[4799]: I1124 07:08:07.899718 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 07:08:07 crc kubenswrapper[4799]: I1124 07:08:07.901718 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 07:08:08 crc kubenswrapper[4799]: I1124 07:08:08.480114 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 07:08:08 crc kubenswrapper[4799]: I1124 07:08:08.607137 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 07:08:08 crc kubenswrapper[4799]: I1124 07:08:08.862888 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-7mn2f"] Nov 24 07:08:08 crc kubenswrapper[4799]: I1124 07:08:08.864804 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:08 crc kubenswrapper[4799]: I1124 07:08:08.899016 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-7mn2f"] Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.004052 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-nb\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.004517 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-swift-storage-0\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.004565 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-sb\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.004674 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-svc\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.004709 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-config\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.004832 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq46q\" (UniqueName: \"kubernetes.io/projected/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-kube-api-access-mq46q\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.106318 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-svc\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.106360 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-config\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.106425 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq46q\" (UniqueName: \"kubernetes.io/projected/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-kube-api-access-mq46q\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.106490 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-nb\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.106521 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-swift-storage-0\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.106543 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-sb\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.107563 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-sb\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.107630 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-config\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.107651 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-nb\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.107921 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-swift-storage-0\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.108351 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-svc\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.134725 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq46q\" (UniqueName: \"kubernetes.io/projected/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-kube-api-access-mq46q\") pod \"dnsmasq-dns-5d7f54fb65-7mn2f\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.204169 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:09 crc kubenswrapper[4799]: I1124 07:08:09.744120 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-7mn2f"] Nov 24 07:08:10 crc kubenswrapper[4799]: I1124 07:08:10.437889 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:10 crc kubenswrapper[4799]: I1124 07:08:10.500691 4799 generic.go:334] "Generic (PLEG): container finished" podID="ce1d3095-b344-4e6c-b49a-d1de3a7260ee" containerID="a566d8f356e5254e7334281ca19afeaafb46a936eff3f95e9bb6c95fec8e862f" exitCode=0 Nov 24 07:08:10 crc kubenswrapper[4799]: I1124 07:08:10.502073 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" event={"ID":"ce1d3095-b344-4e6c-b49a-d1de3a7260ee","Type":"ContainerDied","Data":"a566d8f356e5254e7334281ca19afeaafb46a936eff3f95e9bb6c95fec8e862f"} Nov 24 07:08:10 crc kubenswrapper[4799]: I1124 07:08:10.502106 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" event={"ID":"ce1d3095-b344-4e6c-b49a-d1de3a7260ee","Type":"ContainerStarted","Data":"0f59b71e9d63beedde31ac4aa84625c01cd71aaf89c78e2feb94ca1cbe9291ec"} Nov 24 07:08:10 crc kubenswrapper[4799]: I1124 07:08:10.669185 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:10 crc kubenswrapper[4799]: I1124 07:08:10.669531 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="ceilometer-central-agent" containerID="cri-o://8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd" gracePeriod=30 Nov 24 07:08:10 crc kubenswrapper[4799]: I1124 07:08:10.669581 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="sg-core" containerID="cri-o://f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb" gracePeriod=30 Nov 24 07:08:10 crc kubenswrapper[4799]: I1124 07:08:10.669611 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="ceilometer-notification-agent" containerID="cri-o://7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1" gracePeriod=30 Nov 24 07:08:10 crc kubenswrapper[4799]: I1124 07:08:10.669793 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="proxy-httpd" containerID="cri-o://a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d" gracePeriod=30 Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.511880 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" event={"ID":"ce1d3095-b344-4e6c-b49a-d1de3a7260ee","Type":"ContainerStarted","Data":"38b78508876ab5b361ddd58cb9e13822b959fef1c0bad5d5fe801de1bd871409"} Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.512272 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.516723 4799 generic.go:334] "Generic (PLEG): container finished" podID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerID="a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d" exitCode=0 Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.516763 4799 generic.go:334] "Generic (PLEG): container finished" podID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerID="f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb" exitCode=2 Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.516773 4799 generic.go:334] "Generic (PLEG): container finished" podID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerID="8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd" exitCode=0 Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.516801 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"133ef3af-ba10-4ad0-bee3-dce60aac408f","Type":"ContainerDied","Data":"a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d"} Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.516827 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"133ef3af-ba10-4ad0-bee3-dce60aac408f","Type":"ContainerDied","Data":"f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb"} Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.516836 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"133ef3af-ba10-4ad0-bee3-dce60aac408f","Type":"ContainerDied","Data":"8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd"} Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.540772 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" podStartSLOduration=3.540750232 podStartE2EDuration="3.540750232s" podCreationTimestamp="2025-11-24 07:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:08:11.532281332 +0000 UTC m=+1237.188263806" watchObservedRunningTime="2025-11-24 07:08:11.540750232 +0000 UTC m=+1237.196732706" Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.663501 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.663750 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerName="nova-api-log" containerID="cri-o://8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0" gracePeriod=30 Nov 24 07:08:11 crc kubenswrapper[4799]: I1124 07:08:11.663865 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerName="nova-api-api" containerID="cri-o://24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93" gracePeriod=30 Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.406549 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.479429 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-sg-core-conf-yaml\") pod \"133ef3af-ba10-4ad0-bee3-dce60aac408f\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.479560 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-combined-ca-bundle\") pod \"133ef3af-ba10-4ad0-bee3-dce60aac408f\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.479681 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-config-data\") pod \"133ef3af-ba10-4ad0-bee3-dce60aac408f\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.479709 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-scripts\") pod \"133ef3af-ba10-4ad0-bee3-dce60aac408f\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.479751 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-ceilometer-tls-certs\") pod \"133ef3af-ba10-4ad0-bee3-dce60aac408f\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.479939 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-run-httpd\") pod \"133ef3af-ba10-4ad0-bee3-dce60aac408f\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.479967 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-log-httpd\") pod \"133ef3af-ba10-4ad0-bee3-dce60aac408f\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.480060 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkqkk\" (UniqueName: \"kubernetes.io/projected/133ef3af-ba10-4ad0-bee3-dce60aac408f-kube-api-access-hkqkk\") pod \"133ef3af-ba10-4ad0-bee3-dce60aac408f\" (UID: \"133ef3af-ba10-4ad0-bee3-dce60aac408f\") " Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.480616 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "133ef3af-ba10-4ad0-bee3-dce60aac408f" (UID: "133ef3af-ba10-4ad0-bee3-dce60aac408f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.480775 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "133ef3af-ba10-4ad0-bee3-dce60aac408f" (UID: "133ef3af-ba10-4ad0-bee3-dce60aac408f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.486928 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/133ef3af-ba10-4ad0-bee3-dce60aac408f-kube-api-access-hkqkk" (OuterVolumeSpecName: "kube-api-access-hkqkk") pod "133ef3af-ba10-4ad0-bee3-dce60aac408f" (UID: "133ef3af-ba10-4ad0-bee3-dce60aac408f"). InnerVolumeSpecName "kube-api-access-hkqkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.492642 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-scripts" (OuterVolumeSpecName: "scripts") pod "133ef3af-ba10-4ad0-bee3-dce60aac408f" (UID: "133ef3af-ba10-4ad0-bee3-dce60aac408f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.520278 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "133ef3af-ba10-4ad0-bee3-dce60aac408f" (UID: "133ef3af-ba10-4ad0-bee3-dce60aac408f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.532518 4799 generic.go:334] "Generic (PLEG): container finished" podID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerID="8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0" exitCode=143 Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.532778 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cbed3e4-9342-4cd7-a142-9926ad732782","Type":"ContainerDied","Data":"8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0"} Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.537957 4799 generic.go:334] "Generic (PLEG): container finished" podID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerID="7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1" exitCode=0 Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.538121 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.538040 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"133ef3af-ba10-4ad0-bee3-dce60aac408f","Type":"ContainerDied","Data":"7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1"} Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.538324 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"133ef3af-ba10-4ad0-bee3-dce60aac408f","Type":"ContainerDied","Data":"115bbe449ce97aa5b7ff52dcaea8564db9cacc851c1305ad9cfec244201a7579"} Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.538410 4799 scope.go:117] "RemoveContainer" containerID="a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.585085 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "133ef3af-ba10-4ad0-bee3-dce60aac408f" (UID: "133ef3af-ba10-4ad0-bee3-dce60aac408f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.586961 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.586988 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.587000 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.587009 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/133ef3af-ba10-4ad0-bee3-dce60aac408f-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.587018 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkqkk\" (UniqueName: \"kubernetes.io/projected/133ef3af-ba10-4ad0-bee3-dce60aac408f-kube-api-access-hkqkk\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.587029 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.597228 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "133ef3af-ba10-4ad0-bee3-dce60aac408f" (UID: "133ef3af-ba10-4ad0-bee3-dce60aac408f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.638120 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-config-data" (OuterVolumeSpecName: "config-data") pod "133ef3af-ba10-4ad0-bee3-dce60aac408f" (UID: "133ef3af-ba10-4ad0-bee3-dce60aac408f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.688765 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.689893 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133ef3af-ba10-4ad0-bee3-dce60aac408f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.716168 4799 scope.go:117] "RemoveContainer" containerID="f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.738533 4799 scope.go:117] "RemoveContainer" containerID="7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.769189 4799 scope.go:117] "RemoveContainer" containerID="8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.796807 4799 scope.go:117] "RemoveContainer" containerID="a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d" Nov 24 07:08:12 crc kubenswrapper[4799]: E1124 07:08:12.797475 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d\": container with ID starting with a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d not found: ID does not exist" containerID="a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.797514 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d"} err="failed to get container status \"a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d\": rpc error: code = NotFound desc = could not find container \"a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d\": container with ID starting with a0aab6b8de821687cd075c9ee091312f0285d53c2e7eda9522951d7d4115e51d not found: ID does not exist" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.797545 4799 scope.go:117] "RemoveContainer" containerID="f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb" Nov 24 07:08:12 crc kubenswrapper[4799]: E1124 07:08:12.797979 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb\": container with ID starting with f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb not found: ID does not exist" containerID="f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.798001 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb"} err="failed to get container status \"f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb\": rpc error: code = NotFound desc = could not find container \"f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb\": container with ID starting with f0ecbbd5228f9ca6559f02d759acf3cccd5c2d99d78f50168acc48a35dc71fbb not found: ID does not exist" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.798016 4799 scope.go:117] "RemoveContainer" containerID="7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1" Nov 24 07:08:12 crc kubenswrapper[4799]: E1124 07:08:12.798257 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1\": container with ID starting with 7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1 not found: ID does not exist" containerID="7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.798275 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1"} err="failed to get container status \"7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1\": rpc error: code = NotFound desc = could not find container \"7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1\": container with ID starting with 7b0f47a439eaf2a7fbb615795c13664a8b31d23c56b17bad987474270f015bb1 not found: ID does not exist" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.798289 4799 scope.go:117] "RemoveContainer" containerID="8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd" Nov 24 07:08:12 crc kubenswrapper[4799]: E1124 07:08:12.798665 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd\": container with ID starting with 8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd not found: ID does not exist" containerID="8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.798738 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd"} err="failed to get container status \"8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd\": rpc error: code = NotFound desc = could not find container \"8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd\": container with ID starting with 8d9e74d854ae35401d6b643d510f65764faa3ba7693022a4072fa524ffe5f7fd not found: ID does not exist" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.874342 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.885242 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.906489 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:12 crc kubenswrapper[4799]: E1124 07:08:12.906957 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="ceilometer-central-agent" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.906975 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="ceilometer-central-agent" Nov 24 07:08:12 crc kubenswrapper[4799]: E1124 07:08:12.906984 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="ceilometer-notification-agent" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.906991 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="ceilometer-notification-agent" Nov 24 07:08:12 crc kubenswrapper[4799]: E1124 07:08:12.907003 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="proxy-httpd" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.907009 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="proxy-httpd" Nov 24 07:08:12 crc kubenswrapper[4799]: E1124 07:08:12.907020 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="sg-core" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.907028 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="sg-core" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.907200 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="proxy-httpd" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.907221 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="sg-core" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.907237 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="ceilometer-central-agent" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.907247 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" containerName="ceilometer-notification-agent" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.909044 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.911108 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.912338 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.912382 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.920145 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.997437 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.997501 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blrgn\" (UniqueName: \"kubernetes.io/projected/67db54c5-b187-461e-9dcc-e477266c8341-kube-api-access-blrgn\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.997541 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.997563 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-config-data\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.997641 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.997681 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-scripts\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.997696 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-run-httpd\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:12 crc kubenswrapper[4799]: I1124 07:08:12.997718 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-log-httpd\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.036431 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:13 crc kubenswrapper[4799]: E1124 07:08:13.037234 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-blrgn log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="67db54c5-b187-461e-9dcc-e477266c8341" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.099984 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.100052 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-config-data\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.100175 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.100236 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-scripts\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.100259 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-run-httpd\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.100293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-log-httpd\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.100343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.100379 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blrgn\" (UniqueName: \"kubernetes.io/projected/67db54c5-b187-461e-9dcc-e477266c8341-kube-api-access-blrgn\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.101778 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-log-httpd\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.101785 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-run-httpd\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.106496 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.106579 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-scripts\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.106810 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.106825 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.107446 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-config-data\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.129891 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blrgn\" (UniqueName: \"kubernetes.io/projected/67db54c5-b187-461e-9dcc-e477266c8341-kube-api-access-blrgn\") pod \"ceilometer-0\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.548961 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.560749 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.610138 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-log-httpd\") pod \"67db54c5-b187-461e-9dcc-e477266c8341\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.610590 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-scripts\") pod \"67db54c5-b187-461e-9dcc-e477266c8341\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.610711 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-ceilometer-tls-certs\") pod \"67db54c5-b187-461e-9dcc-e477266c8341\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.610453 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "67db54c5-b187-461e-9dcc-e477266c8341" (UID: "67db54c5-b187-461e-9dcc-e477266c8341"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.611072 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-combined-ca-bundle\") pod \"67db54c5-b187-461e-9dcc-e477266c8341\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.611237 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blrgn\" (UniqueName: \"kubernetes.io/projected/67db54c5-b187-461e-9dcc-e477266c8341-kube-api-access-blrgn\") pod \"67db54c5-b187-461e-9dcc-e477266c8341\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.611464 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-sg-core-conf-yaml\") pod \"67db54c5-b187-461e-9dcc-e477266c8341\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.612298 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-run-httpd\") pod \"67db54c5-b187-461e-9dcc-e477266c8341\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.612596 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-config-data\") pod \"67db54c5-b187-461e-9dcc-e477266c8341\" (UID: \"67db54c5-b187-461e-9dcc-e477266c8341\") " Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.613036 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "67db54c5-b187-461e-9dcc-e477266c8341" (UID: "67db54c5-b187-461e-9dcc-e477266c8341"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.614045 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.614086 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67db54c5-b187-461e-9dcc-e477266c8341-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.616460 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67db54c5-b187-461e-9dcc-e477266c8341-kube-api-access-blrgn" (OuterVolumeSpecName: "kube-api-access-blrgn") pod "67db54c5-b187-461e-9dcc-e477266c8341" (UID: "67db54c5-b187-461e-9dcc-e477266c8341"). InnerVolumeSpecName "kube-api-access-blrgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.618764 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-config-data" (OuterVolumeSpecName: "config-data") pod "67db54c5-b187-461e-9dcc-e477266c8341" (UID: "67db54c5-b187-461e-9dcc-e477266c8341"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.619284 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67db54c5-b187-461e-9dcc-e477266c8341" (UID: "67db54c5-b187-461e-9dcc-e477266c8341"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.618735 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "67db54c5-b187-461e-9dcc-e477266c8341" (UID: "67db54c5-b187-461e-9dcc-e477266c8341"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.629553 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "67db54c5-b187-461e-9dcc-e477266c8341" (UID: "67db54c5-b187-461e-9dcc-e477266c8341"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.638247 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-scripts" (OuterVolumeSpecName: "scripts") pod "67db54c5-b187-461e-9dcc-e477266c8341" (UID: "67db54c5-b187-461e-9dcc-e477266c8341"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.650881 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="133ef3af-ba10-4ad0-bee3-dce60aac408f" path="/var/lib/kubelet/pods/133ef3af-ba10-4ad0-bee3-dce60aac408f/volumes" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.715604 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.715887 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.715992 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.716071 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.716143 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67db54c5-b187-461e-9dcc-e477266c8341-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:13 crc kubenswrapper[4799]: I1124 07:08:13.716218 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blrgn\" (UniqueName: \"kubernetes.io/projected/67db54c5-b187-461e-9dcc-e477266c8341-kube-api-access-blrgn\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.562087 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.649269 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.649340 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.665893 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.669014 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.673255 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.673399 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.674029 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.680952 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.737553 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-run-httpd\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.737626 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.737793 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-scripts\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.737810 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-config-data\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.737864 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.737896 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.737968 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-log-httpd\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.737990 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4xlw\" (UniqueName: \"kubernetes.io/projected/f97d9a56-cb78-4da9-b832-0028b6e45895-kube-api-access-q4xlw\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.840689 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-run-httpd\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.840776 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.840946 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-config-data\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.840979 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-scripts\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.841028 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.841778 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-run-httpd\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.841921 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.842165 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-log-httpd\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.842246 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4xlw\" (UniqueName: \"kubernetes.io/projected/f97d9a56-cb78-4da9-b832-0028b6e45895-kube-api-access-q4xlw\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.842658 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-log-httpd\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.849601 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.849814 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-scripts\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.850248 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-config-data\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.850260 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.850522 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:14 crc kubenswrapper[4799]: I1124 07:08:14.873268 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4xlw\" (UniqueName: \"kubernetes.io/projected/f97d9a56-cb78-4da9-b832-0028b6e45895-kube-api-access-q4xlw\") pod \"ceilometer-0\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " pod="openstack/ceilometer-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.002430 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.294632 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.350776 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-config-data\") pod \"8cbed3e4-9342-4cd7-a142-9926ad732782\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.350870 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqkls\" (UniqueName: \"kubernetes.io/projected/8cbed3e4-9342-4cd7-a142-9926ad732782-kube-api-access-mqkls\") pod \"8cbed3e4-9342-4cd7-a142-9926ad732782\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.350975 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbed3e4-9342-4cd7-a142-9926ad732782-logs\") pod \"8cbed3e4-9342-4cd7-a142-9926ad732782\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.351125 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-combined-ca-bundle\") pod \"8cbed3e4-9342-4cd7-a142-9926ad732782\" (UID: \"8cbed3e4-9342-4cd7-a142-9926ad732782\") " Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.351954 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cbed3e4-9342-4cd7-a142-9926ad732782-logs" (OuterVolumeSpecName: "logs") pod "8cbed3e4-9342-4cd7-a142-9926ad732782" (UID: "8cbed3e4-9342-4cd7-a142-9926ad732782"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.359175 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cbed3e4-9342-4cd7-a142-9926ad732782-kube-api-access-mqkls" (OuterVolumeSpecName: "kube-api-access-mqkls") pod "8cbed3e4-9342-4cd7-a142-9926ad732782" (UID: "8cbed3e4-9342-4cd7-a142-9926ad732782"). InnerVolumeSpecName "kube-api-access-mqkls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.385062 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-config-data" (OuterVolumeSpecName: "config-data") pod "8cbed3e4-9342-4cd7-a142-9926ad732782" (UID: "8cbed3e4-9342-4cd7-a142-9926ad732782"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.390669 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8cbed3e4-9342-4cd7-a142-9926ad732782" (UID: "8cbed3e4-9342-4cd7-a142-9926ad732782"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.436023 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.453279 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.453314 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbed3e4-9342-4cd7-a142-9926ad732782-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.453325 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqkls\" (UniqueName: \"kubernetes.io/projected/8cbed3e4-9342-4cd7-a142-9926ad732782-kube-api-access-mqkls\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.453335 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbed3e4-9342-4cd7-a142-9926ad732782-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.462583 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.517925 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:08:15 crc kubenswrapper[4799]: W1124 07:08:15.521540 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf97d9a56_cb78_4da9_b832_0028b6e45895.slice/crio-1210aa6be7227456168aeae1f4893423c3e858db424c8c026e2f2703c17d5d42 WatchSource:0}: Error finding container 1210aa6be7227456168aeae1f4893423c3e858db424c8c026e2f2703c17d5d42: Status 404 returned error can't find the container with id 1210aa6be7227456168aeae1f4893423c3e858db424c8c026e2f2703c17d5d42 Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.570618 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f97d9a56-cb78-4da9-b832-0028b6e45895","Type":"ContainerStarted","Data":"1210aa6be7227456168aeae1f4893423c3e858db424c8c026e2f2703c17d5d42"} Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.574417 4799 generic.go:334] "Generic (PLEG): container finished" podID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerID="24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93" exitCode=0 Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.574463 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cbed3e4-9342-4cd7-a142-9926ad732782","Type":"ContainerDied","Data":"24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93"} Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.574510 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cbed3e4-9342-4cd7-a142-9926ad732782","Type":"ContainerDied","Data":"5ba62f83eefe7036ffa500cc074774500f001e31e2784b12d9d7358f4a82e1f3"} Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.574533 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.574532 4799 scope.go:117] "RemoveContainer" containerID="24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.610515 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.617745 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.622636 4799 scope.go:117] "RemoveContainer" containerID="8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.666709 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67db54c5-b187-461e-9dcc-e477266c8341" path="/var/lib/kubelet/pods/67db54c5-b187-461e-9dcc-e477266c8341/volumes" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.674168 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.674232 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:15 crc kubenswrapper[4799]: E1124 07:08:15.674748 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerName="nova-api-api" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.674778 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerName="nova-api-api" Nov 24 07:08:15 crc kubenswrapper[4799]: E1124 07:08:15.674830 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerName="nova-api-log" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.683573 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerName="nova-api-log" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.684005 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerName="nova-api-api" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.684039 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbed3e4-9342-4cd7-a142-9926ad732782" containerName="nova-api-log" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.685410 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.693370 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.693758 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.693941 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.694119 4799 scope.go:117] "RemoveContainer" containerID="24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93" Nov 24 07:08:15 crc kubenswrapper[4799]: E1124 07:08:15.695117 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93\": container with ID starting with 24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93 not found: ID does not exist" containerID="24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.695188 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93"} err="failed to get container status \"24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93\": rpc error: code = NotFound desc = could not find container \"24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93\": container with ID starting with 24c6b5cda0634b759a8e0503efffb98c857cd5d695f74605a6bacede5e23bf93 not found: ID does not exist" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.695229 4799 scope.go:117] "RemoveContainer" containerID="8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.695437 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:15 crc kubenswrapper[4799]: E1124 07:08:15.702018 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0\": container with ID starting with 8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0 not found: ID does not exist" containerID="8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.702069 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0"} err="failed to get container status \"8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0\": rpc error: code = NotFound desc = could not find container \"8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0\": container with ID starting with 8d5a7951f6b1ef03476202063bb7a77c09387eaa38ff1efacca7c2344ab7a6d0 not found: ID does not exist" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.761194 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccw82\" (UniqueName: \"kubernetes.io/projected/be229ca4-c9dd-4405-9625-1410f27a25db-kube-api-access-ccw82\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.761552 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be229ca4-c9dd-4405-9625-1410f27a25db-logs\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.761657 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-internal-tls-certs\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.761795 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-public-tls-certs\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.761922 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-config-data\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.762293 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.866507 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be229ca4-c9dd-4405-9625-1410f27a25db-logs\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.866575 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-internal-tls-certs\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.866627 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-public-tls-certs\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.866655 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-config-data\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.866703 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.866769 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccw82\" (UniqueName: \"kubernetes.io/projected/be229ca4-c9dd-4405-9625-1410f27a25db-kube-api-access-ccw82\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.867717 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be229ca4-c9dd-4405-9625-1410f27a25db-logs\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.875197 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-config-data\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.875473 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-internal-tls-certs\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.882183 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.882346 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-public-tls-certs\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.889318 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccw82\" (UniqueName: \"kubernetes.io/projected/be229ca4-c9dd-4405-9625-1410f27a25db-kube-api-access-ccw82\") pod \"nova-api-0\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " pod="openstack/nova-api-0" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.933893 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-x9mds"] Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.935187 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.938813 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.938980 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 07:08:15 crc kubenswrapper[4799]: I1124 07:08:15.951293 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-x9mds"] Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.060401 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.069623 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-config-data\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.069697 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.069757 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2mxz\" (UniqueName: \"kubernetes.io/projected/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-kube-api-access-t2mxz\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.069785 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-scripts\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.171797 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-config-data\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.171890 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.171937 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2mxz\" (UniqueName: \"kubernetes.io/projected/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-kube-api-access-t2mxz\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.171952 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-scripts\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.180227 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-scripts\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.181062 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-config-data\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.190412 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2mxz\" (UniqueName: \"kubernetes.io/projected/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-kube-api-access-t2mxz\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.191766 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-x9mds\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.369101 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.556610 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:16 crc kubenswrapper[4799]: W1124 07:08:16.567151 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe229ca4_c9dd_4405_9625_1410f27a25db.slice/crio-f6da1c34093aa4f57f32686e10772c9281e068fb4159a26a715f100da1516014 WatchSource:0}: Error finding container f6da1c34093aa4f57f32686e10772c9281e068fb4159a26a715f100da1516014: Status 404 returned error can't find the container with id f6da1c34093aa4f57f32686e10772c9281e068fb4159a26a715f100da1516014 Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.596178 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be229ca4-c9dd-4405-9625-1410f27a25db","Type":"ContainerStarted","Data":"f6da1c34093aa4f57f32686e10772c9281e068fb4159a26a715f100da1516014"} Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.609438 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f97d9a56-cb78-4da9-b832-0028b6e45895","Type":"ContainerStarted","Data":"4b1017dff22faa9e7cf62359f3de7ceea284aa7bf591e9d872e15ff649d22d96"} Nov 24 07:08:16 crc kubenswrapper[4799]: I1124 07:08:16.873333 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-x9mds"] Nov 24 07:08:16 crc kubenswrapper[4799]: W1124 07:08:16.877493 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4cd0282_fb2c_48bf_91cb_ae8bd6e17c6d.slice/crio-1ebcef8604235a8f22d075a55ee7f822602f124fd7a362c2b8b21cc5d1e00a26 WatchSource:0}: Error finding container 1ebcef8604235a8f22d075a55ee7f822602f124fd7a362c2b8b21cc5d1e00a26: Status 404 returned error can't find the container with id 1ebcef8604235a8f22d075a55ee7f822602f124fd7a362c2b8b21cc5d1e00a26 Nov 24 07:08:17 crc kubenswrapper[4799]: I1124 07:08:17.624917 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x9mds" event={"ID":"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d","Type":"ContainerStarted","Data":"b59824ec9a02f2da4aecd1a2a4ddb873b0b15ff04f45160a866423c478cd5972"} Nov 24 07:08:17 crc kubenswrapper[4799]: I1124 07:08:17.625357 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x9mds" event={"ID":"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d","Type":"ContainerStarted","Data":"1ebcef8604235a8f22d075a55ee7f822602f124fd7a362c2b8b21cc5d1e00a26"} Nov 24 07:08:17 crc kubenswrapper[4799]: I1124 07:08:17.663734 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cbed3e4-9342-4cd7-a142-9926ad732782" path="/var/lib/kubelet/pods/8cbed3e4-9342-4cd7-a142-9926ad732782/volumes" Nov 24 07:08:17 crc kubenswrapper[4799]: I1124 07:08:17.664038 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-x9mds" podStartSLOduration=2.6640106340000003 podStartE2EDuration="2.664010634s" podCreationTimestamp="2025-11-24 07:08:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:08:17.661518893 +0000 UTC m=+1243.317501367" watchObservedRunningTime="2025-11-24 07:08:17.664010634 +0000 UTC m=+1243.319993138" Nov 24 07:08:17 crc kubenswrapper[4799]: I1124 07:08:17.664954 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be229ca4-c9dd-4405-9625-1410f27a25db","Type":"ContainerStarted","Data":"f49207a0fd4e258a54ba7d5d675c4a1baca4909c64153f44def3ba4a0fdb7648"} Nov 24 07:08:17 crc kubenswrapper[4799]: I1124 07:08:17.665014 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be229ca4-c9dd-4405-9625-1410f27a25db","Type":"ContainerStarted","Data":"89094561e99eefb7f090ddc65af16bee0480b9d9b8bcd4046ee74c11705f4ce8"} Nov 24 07:08:17 crc kubenswrapper[4799]: I1124 07:08:17.665037 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f97d9a56-cb78-4da9-b832-0028b6e45895","Type":"ContainerStarted","Data":"6b464ae1867d289634184c0a2a89b04d292707252e13139c622836b8234cc5e5"} Nov 24 07:08:17 crc kubenswrapper[4799]: I1124 07:08:17.665057 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f97d9a56-cb78-4da9-b832-0028b6e45895","Type":"ContainerStarted","Data":"247ac76988e78afd0fb67b14cc28cf0d059ad2c2f3abbf816f4a4d4fb2204cb3"} Nov 24 07:08:17 crc kubenswrapper[4799]: I1124 07:08:17.700388 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.700361435 podStartE2EDuration="2.700361435s" podCreationTimestamp="2025-11-24 07:08:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:08:17.687747328 +0000 UTC m=+1243.343729842" watchObservedRunningTime="2025-11-24 07:08:17.700361435 +0000 UTC m=+1243.356343909" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.204975 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.281239 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-bk2hb"] Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.281560 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" podUID="f60aa6ed-3836-4943-b93f-9b4f221cb82e" containerName="dnsmasq-dns" containerID="cri-o://c1b436269325ee06b16c543789a8532f1ca29f89d39ac77a6252f8a6b10370f3" gracePeriod=10 Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.503723 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" podUID="f60aa6ed-3836-4943-b93f-9b4f221cb82e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.185:5353: connect: connection refused" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.666768 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f97d9a56-cb78-4da9-b832-0028b6e45895","Type":"ContainerStarted","Data":"dc5fdc4d831816407f510ddb2a4fc3bcb5bbc530983dfb627e98b50af968d2bd"} Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.668603 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.675868 4799 generic.go:334] "Generic (PLEG): container finished" podID="f60aa6ed-3836-4943-b93f-9b4f221cb82e" containerID="c1b436269325ee06b16c543789a8532f1ca29f89d39ac77a6252f8a6b10370f3" exitCode=0 Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.675931 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" event={"ID":"f60aa6ed-3836-4943-b93f-9b4f221cb82e","Type":"ContainerDied","Data":"c1b436269325ee06b16c543789a8532f1ca29f89d39ac77a6252f8a6b10370f3"} Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.796431 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.823670 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.484177974 podStartE2EDuration="5.823637148s" podCreationTimestamp="2025-11-24 07:08:14 +0000 UTC" firstStartedPulling="2025-11-24 07:08:15.525022655 +0000 UTC m=+1241.181005129" lastFinishedPulling="2025-11-24 07:08:18.864481819 +0000 UTC m=+1244.520464303" observedRunningTime="2025-11-24 07:08:19.720267904 +0000 UTC m=+1245.376250378" watchObservedRunningTime="2025-11-24 07:08:19.823637148 +0000 UTC m=+1245.479619622" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.852012 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t845r\" (UniqueName: \"kubernetes.io/projected/f60aa6ed-3836-4943-b93f-9b4f221cb82e-kube-api-access-t845r\") pod \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.852176 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-svc\") pod \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.852220 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-sb\") pod \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.852375 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-config\") pod \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.852400 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-nb\") pod \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.852448 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-swift-storage-0\") pod \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\" (UID: \"f60aa6ed-3836-4943-b93f-9b4f221cb82e\") " Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.859723 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f60aa6ed-3836-4943-b93f-9b4f221cb82e-kube-api-access-t845r" (OuterVolumeSpecName: "kube-api-access-t845r") pod "f60aa6ed-3836-4943-b93f-9b4f221cb82e" (UID: "f60aa6ed-3836-4943-b93f-9b4f221cb82e"). InnerVolumeSpecName "kube-api-access-t845r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.916175 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f60aa6ed-3836-4943-b93f-9b4f221cb82e" (UID: "f60aa6ed-3836-4943-b93f-9b4f221cb82e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.933588 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f60aa6ed-3836-4943-b93f-9b4f221cb82e" (UID: "f60aa6ed-3836-4943-b93f-9b4f221cb82e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.940004 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f60aa6ed-3836-4943-b93f-9b4f221cb82e" (UID: "f60aa6ed-3836-4943-b93f-9b4f221cb82e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.951508 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-config" (OuterVolumeSpecName: "config") pod "f60aa6ed-3836-4943-b93f-9b4f221cb82e" (UID: "f60aa6ed-3836-4943-b93f-9b4f221cb82e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.955084 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.955126 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.955143 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.955155 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t845r\" (UniqueName: \"kubernetes.io/projected/f60aa6ed-3836-4943-b93f-9b4f221cb82e-kube-api-access-t845r\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.955166 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:19 crc kubenswrapper[4799]: I1124 07:08:19.963456 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f60aa6ed-3836-4943-b93f-9b4f221cb82e" (UID: "f60aa6ed-3836-4943-b93f-9b4f221cb82e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:08:20 crc kubenswrapper[4799]: I1124 07:08:20.057456 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f60aa6ed-3836-4943-b93f-9b4f221cb82e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:20 crc kubenswrapper[4799]: I1124 07:08:20.694093 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" event={"ID":"f60aa6ed-3836-4943-b93f-9b4f221cb82e","Type":"ContainerDied","Data":"1730fa91e1072e574816dacd5cc12229d9a79ab889b1d0574c7b90beca763cb5"} Nov 24 07:08:20 crc kubenswrapper[4799]: I1124 07:08:20.694177 4799 scope.go:117] "RemoveContainer" containerID="c1b436269325ee06b16c543789a8532f1ca29f89d39ac77a6252f8a6b10370f3" Nov 24 07:08:20 crc kubenswrapper[4799]: I1124 07:08:20.695817 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-bk2hb" Nov 24 07:08:20 crc kubenswrapper[4799]: I1124 07:08:20.746172 4799 scope.go:117] "RemoveContainer" containerID="b238dc5989e8a0ab0a49951d5190f429b15a6d6522ab9d6c0b546c98d7071513" Nov 24 07:08:20 crc kubenswrapper[4799]: I1124 07:08:20.746576 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-bk2hb"] Nov 24 07:08:20 crc kubenswrapper[4799]: I1124 07:08:20.757088 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-bk2hb"] Nov 24 07:08:21 crc kubenswrapper[4799]: I1124 07:08:21.637377 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f60aa6ed-3836-4943-b93f-9b4f221cb82e" path="/var/lib/kubelet/pods/f60aa6ed-3836-4943-b93f-9b4f221cb82e/volumes" Nov 24 07:08:22 crc kubenswrapper[4799]: I1124 07:08:22.725293 4799 generic.go:334] "Generic (PLEG): container finished" podID="d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d" containerID="b59824ec9a02f2da4aecd1a2a4ddb873b0b15ff04f45160a866423c478cd5972" exitCode=0 Nov 24 07:08:22 crc kubenswrapper[4799]: I1124 07:08:22.725437 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x9mds" event={"ID":"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d","Type":"ContainerDied","Data":"b59824ec9a02f2da4aecd1a2a4ddb873b0b15ff04f45160a866423c478cd5972"} Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.110375 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.161735 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-scripts\") pod \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.161920 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-combined-ca-bundle\") pod \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.162437 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2mxz\" (UniqueName: \"kubernetes.io/projected/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-kube-api-access-t2mxz\") pod \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.162503 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-config-data\") pod \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\" (UID: \"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d\") " Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.167609 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-kube-api-access-t2mxz" (OuterVolumeSpecName: "kube-api-access-t2mxz") pod "d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d" (UID: "d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d"). InnerVolumeSpecName "kube-api-access-t2mxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.167682 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-scripts" (OuterVolumeSpecName: "scripts") pod "d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d" (UID: "d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.200599 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d" (UID: "d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.206659 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-config-data" (OuterVolumeSpecName: "config-data") pod "d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d" (UID: "d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.265446 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.265491 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.265504 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2mxz\" (UniqueName: \"kubernetes.io/projected/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-kube-api-access-t2mxz\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.265513 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.784826 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x9mds" event={"ID":"d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d","Type":"ContainerDied","Data":"1ebcef8604235a8f22d075a55ee7f822602f124fd7a362c2b8b21cc5d1e00a26"} Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.785198 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ebcef8604235a8f22d075a55ee7f822602f124fd7a362c2b8b21cc5d1e00a26" Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.784931 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x9mds" Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.940479 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.941020 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="be229ca4-c9dd-4405-9625-1410f27a25db" containerName="nova-api-log" containerID="cri-o://89094561e99eefb7f090ddc65af16bee0480b9d9b8bcd4046ee74c11705f4ce8" gracePeriod=30 Nov 24 07:08:24 crc kubenswrapper[4799]: I1124 07:08:24.941124 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="be229ca4-c9dd-4405-9625-1410f27a25db" containerName="nova-api-api" containerID="cri-o://f49207a0fd4e258a54ba7d5d675c4a1baca4909c64153f44def3ba4a0fdb7648" gracePeriod=30 Nov 24 07:08:25 crc kubenswrapper[4799]: I1124 07:08:25.018737 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:08:25 crc kubenswrapper[4799]: I1124 07:08:25.019134 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="58bc8e5c-1b62-4f9e-8e0e-83e40a318193" containerName="nova-scheduler-scheduler" containerID="cri-o://81617cd15db8bcf8cda5a02156d69677d41b237ae096f7be2333e6cb47100016" gracePeriod=30 Nov 24 07:08:25 crc kubenswrapper[4799]: I1124 07:08:25.027342 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:08:25 crc kubenswrapper[4799]: I1124 07:08:25.027642 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-log" containerID="cri-o://2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9" gracePeriod=30 Nov 24 07:08:25 crc kubenswrapper[4799]: I1124 07:08:25.027742 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-metadata" containerID="cri-o://9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34" gracePeriod=30 Nov 24 07:08:25 crc kubenswrapper[4799]: I1124 07:08:25.800534 4799 generic.go:334] "Generic (PLEG): container finished" podID="be229ca4-c9dd-4405-9625-1410f27a25db" containerID="f49207a0fd4e258a54ba7d5d675c4a1baca4909c64153f44def3ba4a0fdb7648" exitCode=0 Nov 24 07:08:25 crc kubenswrapper[4799]: I1124 07:08:25.800566 4799 generic.go:334] "Generic (PLEG): container finished" podID="be229ca4-c9dd-4405-9625-1410f27a25db" containerID="89094561e99eefb7f090ddc65af16bee0480b9d9b8bcd4046ee74c11705f4ce8" exitCode=143 Nov 24 07:08:25 crc kubenswrapper[4799]: I1124 07:08:25.800624 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be229ca4-c9dd-4405-9625-1410f27a25db","Type":"ContainerDied","Data":"f49207a0fd4e258a54ba7d5d675c4a1baca4909c64153f44def3ba4a0fdb7648"} Nov 24 07:08:25 crc kubenswrapper[4799]: I1124 07:08:25.800661 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be229ca4-c9dd-4405-9625-1410f27a25db","Type":"ContainerDied","Data":"89094561e99eefb7f090ddc65af16bee0480b9d9b8bcd4046ee74c11705f4ce8"} Nov 24 07:08:25 crc kubenswrapper[4799]: I1124 07:08:25.811292 4799 generic.go:334] "Generic (PLEG): container finished" podID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerID="2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9" exitCode=143 Nov 24 07:08:25 crc kubenswrapper[4799]: I1124 07:08:25.811349 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bce44781-5a13-4b7c-a9f3-022fcf589dcb","Type":"ContainerDied","Data":"2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9"} Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.104589 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.203912 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-public-tls-certs\") pod \"be229ca4-c9dd-4405-9625-1410f27a25db\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.203977 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-combined-ca-bundle\") pod \"be229ca4-c9dd-4405-9625-1410f27a25db\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.204123 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccw82\" (UniqueName: \"kubernetes.io/projected/be229ca4-c9dd-4405-9625-1410f27a25db-kube-api-access-ccw82\") pod \"be229ca4-c9dd-4405-9625-1410f27a25db\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.204158 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be229ca4-c9dd-4405-9625-1410f27a25db-logs\") pod \"be229ca4-c9dd-4405-9625-1410f27a25db\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.204205 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-config-data\") pod \"be229ca4-c9dd-4405-9625-1410f27a25db\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.204241 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-internal-tls-certs\") pod \"be229ca4-c9dd-4405-9625-1410f27a25db\" (UID: \"be229ca4-c9dd-4405-9625-1410f27a25db\") " Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.205246 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be229ca4-c9dd-4405-9625-1410f27a25db-logs" (OuterVolumeSpecName: "logs") pod "be229ca4-c9dd-4405-9625-1410f27a25db" (UID: "be229ca4-c9dd-4405-9625-1410f27a25db"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.216221 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be229ca4-c9dd-4405-9625-1410f27a25db-kube-api-access-ccw82" (OuterVolumeSpecName: "kube-api-access-ccw82") pod "be229ca4-c9dd-4405-9625-1410f27a25db" (UID: "be229ca4-c9dd-4405-9625-1410f27a25db"). InnerVolumeSpecName "kube-api-access-ccw82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.242665 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-config-data" (OuterVolumeSpecName: "config-data") pod "be229ca4-c9dd-4405-9625-1410f27a25db" (UID: "be229ca4-c9dd-4405-9625-1410f27a25db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.245374 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be229ca4-c9dd-4405-9625-1410f27a25db" (UID: "be229ca4-c9dd-4405-9625-1410f27a25db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.273018 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "be229ca4-c9dd-4405-9625-1410f27a25db" (UID: "be229ca4-c9dd-4405-9625-1410f27a25db"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.275413 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "be229ca4-c9dd-4405-9625-1410f27a25db" (UID: "be229ca4-c9dd-4405-9625-1410f27a25db"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.307307 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccw82\" (UniqueName: \"kubernetes.io/projected/be229ca4-c9dd-4405-9625-1410f27a25db-kube-api-access-ccw82\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.307380 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be229ca4-c9dd-4405-9625-1410f27a25db-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.307397 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.307411 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.307427 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.307442 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be229ca4-c9dd-4405-9625-1410f27a25db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.823632 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"be229ca4-c9dd-4405-9625-1410f27a25db","Type":"ContainerDied","Data":"f6da1c34093aa4f57f32686e10772c9281e068fb4159a26a715f100da1516014"} Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.824126 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.824200 4799 scope.go:117] "RemoveContainer" containerID="f49207a0fd4e258a54ba7d5d675c4a1baca4909c64153f44def3ba4a0fdb7648" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.825747 4799 generic.go:334] "Generic (PLEG): container finished" podID="58bc8e5c-1b62-4f9e-8e0e-83e40a318193" containerID="81617cd15db8bcf8cda5a02156d69677d41b237ae096f7be2333e6cb47100016" exitCode=0 Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.825786 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58bc8e5c-1b62-4f9e-8e0e-83e40a318193","Type":"ContainerDied","Data":"81617cd15db8bcf8cda5a02156d69677d41b237ae096f7be2333e6cb47100016"} Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.883931 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.894607 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.900675 4799 scope.go:117] "RemoveContainer" containerID="89094561e99eefb7f090ddc65af16bee0480b9d9b8bcd4046ee74c11705f4ce8" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.925999 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:26 crc kubenswrapper[4799]: E1124 07:08:26.927218 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f60aa6ed-3836-4943-b93f-9b4f221cb82e" containerName="dnsmasq-dns" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.927243 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f60aa6ed-3836-4943-b93f-9b4f221cb82e" containerName="dnsmasq-dns" Nov 24 07:08:26 crc kubenswrapper[4799]: E1124 07:08:26.927254 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d" containerName="nova-manage" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.927261 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d" containerName="nova-manage" Nov 24 07:08:26 crc kubenswrapper[4799]: E1124 07:08:26.927289 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f60aa6ed-3836-4943-b93f-9b4f221cb82e" containerName="init" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.927296 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f60aa6ed-3836-4943-b93f-9b4f221cb82e" containerName="init" Nov 24 07:08:26 crc kubenswrapper[4799]: E1124 07:08:26.927318 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be229ca4-c9dd-4405-9625-1410f27a25db" containerName="nova-api-api" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.927325 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="be229ca4-c9dd-4405-9625-1410f27a25db" containerName="nova-api-api" Nov 24 07:08:26 crc kubenswrapper[4799]: E1124 07:08:26.927346 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be229ca4-c9dd-4405-9625-1410f27a25db" containerName="nova-api-log" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.927352 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="be229ca4-c9dd-4405-9625-1410f27a25db" containerName="nova-api-log" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.927569 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d" containerName="nova-manage" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.927597 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="be229ca4-c9dd-4405-9625-1410f27a25db" containerName="nova-api-api" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.927616 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="be229ca4-c9dd-4405-9625-1410f27a25db" containerName="nova-api-log" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.927628 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f60aa6ed-3836-4943-b93f-9b4f221cb82e" containerName="dnsmasq-dns" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.929916 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.933252 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.933308 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.933328 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 07:08:26 crc kubenswrapper[4799]: I1124 07:08:26.939564 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.027163 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.027237 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.027262 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-public-tls-certs\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.027328 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fe03a76-2eb9-4728-9396-9d2dc1e743af-logs\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.027383 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqksl\" (UniqueName: \"kubernetes.io/projected/6fe03a76-2eb9-4728-9396-9d2dc1e743af-kube-api-access-nqksl\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.027409 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-config-data\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.042554 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.130456 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lhqp\" (UniqueName: \"kubernetes.io/projected/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-kube-api-access-6lhqp\") pod \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.130596 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-combined-ca-bundle\") pod \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.130641 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-config-data\") pod \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\" (UID: \"58bc8e5c-1b62-4f9e-8e0e-83e40a318193\") " Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.131640 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.131707 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.131751 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-public-tls-certs\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.131838 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fe03a76-2eb9-4728-9396-9d2dc1e743af-logs\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.131929 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqksl\" (UniqueName: \"kubernetes.io/projected/6fe03a76-2eb9-4728-9396-9d2dc1e743af-kube-api-access-nqksl\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.131982 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-config-data\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.141036 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-config-data\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.141083 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.144576 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-kube-api-access-6lhqp" (OuterVolumeSpecName: "kube-api-access-6lhqp") pod "58bc8e5c-1b62-4f9e-8e0e-83e40a318193" (UID: "58bc8e5c-1b62-4f9e-8e0e-83e40a318193"). InnerVolumeSpecName "kube-api-access-6lhqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.145548 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.147083 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fe03a76-2eb9-4728-9396-9d2dc1e743af-logs\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.148836 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-public-tls-certs\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.180683 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqksl\" (UniqueName: \"kubernetes.io/projected/6fe03a76-2eb9-4728-9396-9d2dc1e743af-kube-api-access-nqksl\") pod \"nova-api-0\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.181580 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58bc8e5c-1b62-4f9e-8e0e-83e40a318193" (UID: "58bc8e5c-1b62-4f9e-8e0e-83e40a318193"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.200022 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-config-data" (OuterVolumeSpecName: "config-data") pod "58bc8e5c-1b62-4f9e-8e0e-83e40a318193" (UID: "58bc8e5c-1b62-4f9e-8e0e-83e40a318193"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.234621 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lhqp\" (UniqueName: \"kubernetes.io/projected/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-kube-api-access-6lhqp\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.234814 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.234948 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58bc8e5c-1b62-4f9e-8e0e-83e40a318193-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.270401 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.646591 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be229ca4-c9dd-4405-9625-1410f27a25db" path="/var/lib/kubelet/pods/be229ca4-c9dd-4405-9625-1410f27a25db/volumes" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.732229 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:08:27 crc kubenswrapper[4799]: W1124 07:08:27.737509 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fe03a76_2eb9_4728_9396_9d2dc1e743af.slice/crio-482919ed95ee1f06b013f41ee9fc371c03d41adacb105440bc45b712fae39959 WatchSource:0}: Error finding container 482919ed95ee1f06b013f41ee9fc371c03d41adacb105440bc45b712fae39959: Status 404 returned error can't find the container with id 482919ed95ee1f06b013f41ee9fc371c03d41adacb105440bc45b712fae39959 Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.840213 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58bc8e5c-1b62-4f9e-8e0e-83e40a318193","Type":"ContainerDied","Data":"59d2ba4f8b07df33597c9fc41405b43cda395c0d41e140e07c40d1735d7d587e"} Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.840255 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.840267 4799 scope.go:117] "RemoveContainer" containerID="81617cd15db8bcf8cda5a02156d69677d41b237ae096f7be2333e6cb47100016" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.842684 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fe03a76-2eb9-4728-9396-9d2dc1e743af","Type":"ContainerStarted","Data":"482919ed95ee1f06b013f41ee9fc371c03d41adacb105440bc45b712fae39959"} Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.872121 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.889130 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.908628 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:08:27 crc kubenswrapper[4799]: E1124 07:08:27.909077 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58bc8e5c-1b62-4f9e-8e0e-83e40a318193" containerName="nova-scheduler-scheduler" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.909092 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="58bc8e5c-1b62-4f9e-8e0e-83e40a318193" containerName="nova-scheduler-scheduler" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.909408 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="58bc8e5c-1b62-4f9e-8e0e-83e40a318193" containerName="nova-scheduler-scheduler" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.909991 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.910062 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.920376 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.951138 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-config-data\") pod \"nova-scheduler-0\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " pod="openstack/nova-scheduler-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.951203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w69r4\" (UniqueName: \"kubernetes.io/projected/0d55de08-d6b7-44f6-8e96-44fcd8999000-kube-api-access-w69r4\") pod \"nova-scheduler-0\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " pod="openstack/nova-scheduler-0" Nov 24 07:08:27 crc kubenswrapper[4799]: I1124 07:08:27.951389 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " pod="openstack/nova-scheduler-0" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.052321 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " pod="openstack/nova-scheduler-0" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.052392 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-config-data\") pod \"nova-scheduler-0\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " pod="openstack/nova-scheduler-0" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.052427 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w69r4\" (UniqueName: \"kubernetes.io/projected/0d55de08-d6b7-44f6-8e96-44fcd8999000-kube-api-access-w69r4\") pod \"nova-scheduler-0\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " pod="openstack/nova-scheduler-0" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.056068 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-config-data\") pod \"nova-scheduler-0\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " pod="openstack/nova-scheduler-0" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.057445 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " pod="openstack/nova-scheduler-0" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.075585 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w69r4\" (UniqueName: \"kubernetes.io/projected/0d55de08-d6b7-44f6-8e96-44fcd8999000-kube-api-access-w69r4\") pod \"nova-scheduler-0\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " pod="openstack/nova-scheduler-0" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.156145 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:40106->10.217.0.191:8775: read: connection reset by peer" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.156191 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:40110->10.217.0.191:8775: read: connection reset by peer" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.237759 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.584745 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.662864 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz6ng\" (UniqueName: \"kubernetes.io/projected/bce44781-5a13-4b7c-a9f3-022fcf589dcb-kube-api-access-lz6ng\") pod \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.662933 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-nova-metadata-tls-certs\") pod \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.662987 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bce44781-5a13-4b7c-a9f3-022fcf589dcb-logs\") pod \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.664087 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-config-data\") pod \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.664157 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-combined-ca-bundle\") pod \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\" (UID: \"bce44781-5a13-4b7c-a9f3-022fcf589dcb\") " Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.667202 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bce44781-5a13-4b7c-a9f3-022fcf589dcb-logs" (OuterVolumeSpecName: "logs") pod "bce44781-5a13-4b7c-a9f3-022fcf589dcb" (UID: "bce44781-5a13-4b7c-a9f3-022fcf589dcb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.673524 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bce44781-5a13-4b7c-a9f3-022fcf589dcb-kube-api-access-lz6ng" (OuterVolumeSpecName: "kube-api-access-lz6ng") pod "bce44781-5a13-4b7c-a9f3-022fcf589dcb" (UID: "bce44781-5a13-4b7c-a9f3-022fcf589dcb"). InnerVolumeSpecName "kube-api-access-lz6ng". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.696093 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-config-data" (OuterVolumeSpecName: "config-data") pod "bce44781-5a13-4b7c-a9f3-022fcf589dcb" (UID: "bce44781-5a13-4b7c-a9f3-022fcf589dcb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.705164 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bce44781-5a13-4b7c-a9f3-022fcf589dcb" (UID: "bce44781-5a13-4b7c-a9f3-022fcf589dcb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.724490 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "bce44781-5a13-4b7c-a9f3-022fcf589dcb" (UID: "bce44781-5a13-4b7c-a9f3-022fcf589dcb"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.766344 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz6ng\" (UniqueName: \"kubernetes.io/projected/bce44781-5a13-4b7c-a9f3-022fcf589dcb-kube-api-access-lz6ng\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.766390 4799 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.766404 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bce44781-5a13-4b7c-a9f3-022fcf589dcb-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.766413 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.766423 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce44781-5a13-4b7c-a9f3-022fcf589dcb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.844944 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:08:28 crc kubenswrapper[4799]: W1124 07:08:28.853281 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d55de08_d6b7_44f6_8e96_44fcd8999000.slice/crio-97c4e8b888ca0aacccacf821815670ffc6cf28575c35f660f18b84923156e041 WatchSource:0}: Error finding container 97c4e8b888ca0aacccacf821815670ffc6cf28575c35f660f18b84923156e041: Status 404 returned error can't find the container with id 97c4e8b888ca0aacccacf821815670ffc6cf28575c35f660f18b84923156e041 Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.866630 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fe03a76-2eb9-4728-9396-9d2dc1e743af","Type":"ContainerStarted","Data":"f6b04750444de5a3182982a1b53fc81a035e9a9f2eb6df3d21f831648f26fc49"} Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.866675 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fe03a76-2eb9-4728-9396-9d2dc1e743af","Type":"ContainerStarted","Data":"cbf304821538247ce0d3cfcbea062c7ba81f6603d05b995bc8ebc90500583b62"} Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.871201 4799 generic.go:334] "Generic (PLEG): container finished" podID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerID="9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34" exitCode=0 Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.871246 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bce44781-5a13-4b7c-a9f3-022fcf589dcb","Type":"ContainerDied","Data":"9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34"} Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.871271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bce44781-5a13-4b7c-a9f3-022fcf589dcb","Type":"ContainerDied","Data":"d178da4eb487827fa15abb19b96ae96fe9f1d97c281fe32a317d6075674a7b28"} Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.871304 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.871321 4799 scope.go:117] "RemoveContainer" containerID="9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.905029 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.904992213 podStartE2EDuration="2.904992213s" podCreationTimestamp="2025-11-24 07:08:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:08:28.884299405 +0000 UTC m=+1254.540281879" watchObservedRunningTime="2025-11-24 07:08:28.904992213 +0000 UTC m=+1254.560974687" Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.921993 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.934368 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:08:28 crc kubenswrapper[4799]: I1124 07:08:28.947296 4799 scope.go:117] "RemoveContainer" containerID="2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:28.952891 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:08:29 crc kubenswrapper[4799]: E1124 07:08:28.953295 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-metadata" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:28.953309 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-metadata" Nov 24 07:08:29 crc kubenswrapper[4799]: E1124 07:08:28.953328 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-log" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:28.953335 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-log" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:28.953539 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-log" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:28.953559 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" containerName="nova-metadata-metadata" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.000918 4799 scope.go:117] "RemoveContainer" containerID="9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.002296 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.006534 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.006583 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.006623 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-config-data\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.006646 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6218522b-3df9-45b0-8edb-fb935a7f4109-logs\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.006665 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkwpb\" (UniqueName: \"kubernetes.io/projected/6218522b-3df9-45b0-8edb-fb935a7f4109-kube-api-access-vkwpb\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: E1124 07:08:29.008441 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34\": container with ID starting with 9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34 not found: ID does not exist" containerID="9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.008485 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34"} err="failed to get container status \"9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34\": rpc error: code = NotFound desc = could not find container \"9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34\": container with ID starting with 9df6034b6aae0addb15c4a3f3e5ebf8f6286cc765e763526543d8b97b21bff34 not found: ID does not exist" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.008516 4799 scope.go:117] "RemoveContainer" containerID="2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9" Nov 24 07:08:29 crc kubenswrapper[4799]: E1124 07:08:29.010661 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9\": container with ID starting with 2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9 not found: ID does not exist" containerID="2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.010739 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9"} err="failed to get container status \"2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9\": rpc error: code = NotFound desc = could not find container \"2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9\": container with ID starting with 2c0e33de486157aac7b0ae884275a5fc4b3c7f35c3ed1c0399e05307b6ecf3a9 not found: ID does not exist" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.012291 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.012549 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.013760 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.111327 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-config-data\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.111379 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6218522b-3df9-45b0-8edb-fb935a7f4109-logs\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.111410 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkwpb\" (UniqueName: \"kubernetes.io/projected/6218522b-3df9-45b0-8edb-fb935a7f4109-kube-api-access-vkwpb\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.111607 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.111648 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.112670 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6218522b-3df9-45b0-8edb-fb935a7f4109-logs\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.118104 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.118970 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.119369 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-config-data\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.131276 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkwpb\" (UniqueName: \"kubernetes.io/projected/6218522b-3df9-45b0-8edb-fb935a7f4109-kube-api-access-vkwpb\") pod \"nova-metadata-0\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.248046 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.640920 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58bc8e5c-1b62-4f9e-8e0e-83e40a318193" path="/var/lib/kubelet/pods/58bc8e5c-1b62-4f9e-8e0e-83e40a318193/volumes" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.641695 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bce44781-5a13-4b7c-a9f3-022fcf589dcb" path="/var/lib/kubelet/pods/bce44781-5a13-4b7c-a9f3-022fcf589dcb/volumes" Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.849039 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:08:29 crc kubenswrapper[4799]: W1124 07:08:29.855987 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6218522b_3df9_45b0_8edb_fb935a7f4109.slice/crio-a0fadf5e8fdde0305f30ee9718c93de80bc59c9bf6e5f5cb0dc63c81954d2054 WatchSource:0}: Error finding container a0fadf5e8fdde0305f30ee9718c93de80bc59c9bf6e5f5cb0dc63c81954d2054: Status 404 returned error can't find the container with id a0fadf5e8fdde0305f30ee9718c93de80bc59c9bf6e5f5cb0dc63c81954d2054 Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.892219 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6218522b-3df9-45b0-8edb-fb935a7f4109","Type":"ContainerStarted","Data":"a0fadf5e8fdde0305f30ee9718c93de80bc59c9bf6e5f5cb0dc63c81954d2054"} Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.894230 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d55de08-d6b7-44f6-8e96-44fcd8999000","Type":"ContainerStarted","Data":"0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac"} Nov 24 07:08:29 crc kubenswrapper[4799]: I1124 07:08:29.894268 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d55de08-d6b7-44f6-8e96-44fcd8999000","Type":"ContainerStarted","Data":"97c4e8b888ca0aacccacf821815670ffc6cf28575c35f660f18b84923156e041"} Nov 24 07:08:30 crc kubenswrapper[4799]: I1124 07:08:30.915836 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6218522b-3df9-45b0-8edb-fb935a7f4109","Type":"ContainerStarted","Data":"5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61"} Nov 24 07:08:30 crc kubenswrapper[4799]: I1124 07:08:30.916268 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6218522b-3df9-45b0-8edb-fb935a7f4109","Type":"ContainerStarted","Data":"c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547"} Nov 24 07:08:30 crc kubenswrapper[4799]: I1124 07:08:30.971366 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.971328949 podStartE2EDuration="2.971328949s" podCreationTimestamp="2025-11-24 07:08:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:08:30.945094085 +0000 UTC m=+1256.601076619" watchObservedRunningTime="2025-11-24 07:08:30.971328949 +0000 UTC m=+1256.627311453" Nov 24 07:08:30 crc kubenswrapper[4799]: I1124 07:08:30.976055 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.976039403 podStartE2EDuration="3.976039403s" podCreationTimestamp="2025-11-24 07:08:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:08:29.919524871 +0000 UTC m=+1255.575507355" watchObservedRunningTime="2025-11-24 07:08:30.976039403 +0000 UTC m=+1256.632021887" Nov 24 07:08:33 crc kubenswrapper[4799]: I1124 07:08:33.238213 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 07:08:34 crc kubenswrapper[4799]: I1124 07:08:34.249187 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:08:34 crc kubenswrapper[4799]: I1124 07:08:34.249646 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 07:08:37 crc kubenswrapper[4799]: I1124 07:08:37.271633 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:08:37 crc kubenswrapper[4799]: I1124 07:08:37.272232 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 07:08:38 crc kubenswrapper[4799]: I1124 07:08:38.238545 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 07:08:38 crc kubenswrapper[4799]: I1124 07:08:38.279138 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 07:08:38 crc kubenswrapper[4799]: I1124 07:08:38.312158 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:08:38 crc kubenswrapper[4799]: I1124 07:08:38.321144 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:08:39 crc kubenswrapper[4799]: I1124 07:08:39.059513 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 07:08:39 crc kubenswrapper[4799]: I1124 07:08:39.249614 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 07:08:39 crc kubenswrapper[4799]: I1124 07:08:39.249687 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 07:08:40 crc kubenswrapper[4799]: I1124 07:08:40.272035 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:08:40 crc kubenswrapper[4799]: I1124 07:08:40.272222 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 07:08:45 crc kubenswrapper[4799]: I1124 07:08:45.018238 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 07:08:47 crc kubenswrapper[4799]: I1124 07:08:47.280361 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 07:08:47 crc kubenswrapper[4799]: I1124 07:08:47.281459 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 07:08:47 crc kubenswrapper[4799]: I1124 07:08:47.285421 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 07:08:47 crc kubenswrapper[4799]: I1124 07:08:47.291334 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 07:08:48 crc kubenswrapper[4799]: I1124 07:08:48.138105 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 07:08:48 crc kubenswrapper[4799]: I1124 07:08:48.143376 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 07:08:49 crc kubenswrapper[4799]: I1124 07:08:49.256113 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 07:08:49 crc kubenswrapper[4799]: I1124 07:08:49.260292 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 07:08:49 crc kubenswrapper[4799]: I1124 07:08:49.263293 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 07:08:50 crc kubenswrapper[4799]: I1124 07:08:50.166707 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 07:08:50 crc kubenswrapper[4799]: I1124 07:08:50.400710 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:08:50 crc kubenswrapper[4799]: I1124 07:08:50.400790 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.537052 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.537629 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="76771b43-063e-4bc4-a366-3a13f27e3124" containerName="openstackclient" containerID="cri-o://1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259" gracePeriod=2 Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.557297 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.822074 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.841599 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinderbbb1-account-delete-tjfp2"] Nov 24 07:09:09 crc kubenswrapper[4799]: E1124 07:09:09.841987 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76771b43-063e-4bc4-a366-3a13f27e3124" containerName="openstackclient" Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.842003 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="76771b43-063e-4bc4-a366-3a13f27e3124" containerName="openstackclient" Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.842376 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="76771b43-063e-4bc4-a366-3a13f27e3124" containerName="openstackclient" Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.853193 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderbbb1-account-delete-tjfp2" Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.880757 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9064d19-6806-4e78-beb4-cda924fe3010-operator-scripts\") pod \"cinderbbb1-account-delete-tjfp2\" (UID: \"d9064d19-6806-4e78-beb4-cda924fe3010\") " pod="openstack/cinderbbb1-account-delete-tjfp2" Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.880864 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx9gx\" (UniqueName: \"kubernetes.io/projected/d9064d19-6806-4e78-beb4-cda924fe3010-kube-api-access-wx9gx\") pod \"cinderbbb1-account-delete-tjfp2\" (UID: \"d9064d19-6806-4e78-beb4-cda924fe3010\") " pod="openstack/cinderbbb1-account-delete-tjfp2" Nov 24 07:09:09 crc kubenswrapper[4799]: E1124 07:09:09.881447 4799 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 07:09:09 crc kubenswrapper[4799]: E1124 07:09:09.881493 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data podName:96acf88b-3fe8-46ac-8393-37f720632dd6 nodeName:}" failed. No retries permitted until 2025-11-24 07:09:10.381477465 +0000 UTC m=+1296.037459939 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data") pod "rabbitmq-server-0" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6") : configmap "rabbitmq-config-data" not found Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.891708 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinderbbb1-account-delete-tjfp2"] Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.949035 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.949263 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerName="ovn-northd" containerID="cri-o://2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1" gracePeriod=30 Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.949675 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerName="openstack-network-exporter" containerID="cri-o://8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80" gracePeriod=30 Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.968795 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance00d0-account-delete-kbtpl"] Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.975158 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance00d0-account-delete-kbtpl" Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.990937 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq2qg\" (UniqueName: \"kubernetes.io/projected/ad75f933-ab36-4c9d-857d-c2cc11702dd3-kube-api-access-hq2qg\") pod \"glance00d0-account-delete-kbtpl\" (UID: \"ad75f933-ab36-4c9d-857d-c2cc11702dd3\") " pod="openstack/glance00d0-account-delete-kbtpl" Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.991019 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad75f933-ab36-4c9d-857d-c2cc11702dd3-operator-scripts\") pod \"glance00d0-account-delete-kbtpl\" (UID: \"ad75f933-ab36-4c9d-857d-c2cc11702dd3\") " pod="openstack/glance00d0-account-delete-kbtpl" Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.991050 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9064d19-6806-4e78-beb4-cda924fe3010-operator-scripts\") pod \"cinderbbb1-account-delete-tjfp2\" (UID: \"d9064d19-6806-4e78-beb4-cda924fe3010\") " pod="openstack/cinderbbb1-account-delete-tjfp2" Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.991212 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx9gx\" (UniqueName: \"kubernetes.io/projected/d9064d19-6806-4e78-beb4-cda924fe3010-kube-api-access-wx9gx\") pod \"cinderbbb1-account-delete-tjfp2\" (UID: \"d9064d19-6806-4e78-beb4-cda924fe3010\") " pod="openstack/cinderbbb1-account-delete-tjfp2" Nov 24 07:09:09 crc kubenswrapper[4799]: I1124 07:09:09.992278 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9064d19-6806-4e78-beb4-cda924fe3010-operator-scripts\") pod \"cinderbbb1-account-delete-tjfp2\" (UID: \"d9064d19-6806-4e78-beb4-cda924fe3010\") " pod="openstack/cinderbbb1-account-delete-tjfp2" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.011990 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-qfbzk"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.044916 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance00d0-account-delete-kbtpl"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.073593 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx9gx\" (UniqueName: \"kubernetes.io/projected/d9064d19-6806-4e78-beb4-cda924fe3010-kube-api-access-wx9gx\") pod \"cinderbbb1-account-delete-tjfp2\" (UID: \"d9064d19-6806-4e78-beb4-cda924fe3010\") " pod="openstack/cinderbbb1-account-delete-tjfp2" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.075080 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-qfbzk"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.115919 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq2qg\" (UniqueName: \"kubernetes.io/projected/ad75f933-ab36-4c9d-857d-c2cc11702dd3-kube-api-access-hq2qg\") pod \"glance00d0-account-delete-kbtpl\" (UID: \"ad75f933-ab36-4c9d-857d-c2cc11702dd3\") " pod="openstack/glance00d0-account-delete-kbtpl" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.115991 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad75f933-ab36-4c9d-857d-c2cc11702dd3-operator-scripts\") pod \"glance00d0-account-delete-kbtpl\" (UID: \"ad75f933-ab36-4c9d-857d-c2cc11702dd3\") " pod="openstack/glance00d0-account-delete-kbtpl" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.117155 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad75f933-ab36-4c9d-857d-c2cc11702dd3-operator-scripts\") pod \"glance00d0-account-delete-kbtpl\" (UID: \"ad75f933-ab36-4c9d-857d-c2cc11702dd3\") " pod="openstack/glance00d0-account-delete-kbtpl" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.180955 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderbbb1-account-delete-tjfp2" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.188054 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.188608 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="bbfb60f9-6e71-4b22-9d74-cab607a00c20" containerName="openstack-network-exporter" containerID="cri-o://a9ca7d86df080d4b606da7e5a800c32f1929eb3f6ccde8a25eb20a206bdf9fda" gracePeriod=300 Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.188822 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq2qg\" (UniqueName: \"kubernetes.io/projected/ad75f933-ab36-4c9d-857d-c2cc11702dd3-kube-api-access-hq2qg\") pod \"glance00d0-account-delete-kbtpl\" (UID: \"ad75f933-ab36-4c9d-857d-c2cc11702dd3\") " pod="openstack/glance00d0-account-delete-kbtpl" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.317893 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.320233 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance00d0-account-delete-kbtpl" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.338954 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron2d2f-account-delete-5c2sd"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.340221 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron2d2f-account-delete-5c2sd" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.369365 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-operator-scripts\") pod \"neutron2d2f-account-delete-5c2sd\" (UID: \"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee\") " pod="openstack/neutron2d2f-account-delete-5c2sd" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.369452 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zn9g\" (UniqueName: \"kubernetes.io/projected/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-kube-api-access-8zn9g\") pod \"neutron2d2f-account-delete-5c2sd\" (UID: \"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee\") " pod="openstack/neutron2d2f-account-delete-5c2sd" Nov 24 07:09:10 crc kubenswrapper[4799]: E1124 07:09:10.371186 4799 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 07:09:10 crc kubenswrapper[4799]: E1124 07:09:10.371226 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data podName:366b7548-2fca-4623-a45c-c1e0367ce93a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:10.871213312 +0000 UTC m=+1296.527195786 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data") pod "rabbitmq-cell1-server-0" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a") : configmap "rabbitmq-cell1-config-data" not found Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.376463 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron2d2f-account-delete-5c2sd"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.386174 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican44f2-account-delete-2hm9q"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.387453 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican44f2-account-delete-2hm9q" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.404439 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican44f2-account-delete-2hm9q"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.423730 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-5wg47"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.431529 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-5wg47"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.441745 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement2704-account-delete-ck9t2"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.444965 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement2704-account-delete-ck9t2" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.458272 4799 generic.go:334] "Generic (PLEG): container finished" podID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerID="8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80" exitCode=2 Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.458359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6aa59428-ee60-4434-9245-3ff8fa55200e","Type":"ContainerDied","Data":"8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80"} Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.480337 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4cdf\" (UniqueName: \"kubernetes.io/projected/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-kube-api-access-c4cdf\") pod \"barbican44f2-account-delete-2hm9q\" (UID: \"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a\") " pod="openstack/barbican44f2-account-delete-2hm9q" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.480411 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts\") pod \"barbican44f2-account-delete-2hm9q\" (UID: \"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a\") " pod="openstack/barbican44f2-account-delete-2hm9q" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.480490 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-operator-scripts\") pod \"neutron2d2f-account-delete-5c2sd\" (UID: \"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee\") " pod="openstack/neutron2d2f-account-delete-5c2sd" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.480562 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zn9g\" (UniqueName: \"kubernetes.io/projected/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-kube-api-access-8zn9g\") pod \"neutron2d2f-account-delete-5c2sd\" (UID: \"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee\") " pod="openstack/neutron2d2f-account-delete-5c2sd" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.480595 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nkk7\" (UniqueName: \"kubernetes.io/projected/b6f22313-6d9a-4b35-b892-695903d354b1-kube-api-access-8nkk7\") pod \"placement2704-account-delete-ck9t2\" (UID: \"b6f22313-6d9a-4b35-b892-695903d354b1\") " pod="openstack/placement2704-account-delete-ck9t2" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.480676 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts\") pod \"placement2704-account-delete-ck9t2\" (UID: \"b6f22313-6d9a-4b35-b892-695903d354b1\") " pod="openstack/placement2704-account-delete-ck9t2" Nov 24 07:09:10 crc kubenswrapper[4799]: E1124 07:09:10.480824 4799 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 07:09:10 crc kubenswrapper[4799]: E1124 07:09:10.480891 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data podName:96acf88b-3fe8-46ac-8393-37f720632dd6 nodeName:}" failed. No retries permitted until 2025-11-24 07:09:11.480873954 +0000 UTC m=+1297.136856428 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data") pod "rabbitmq-server-0" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6") : configmap "rabbitmq-config-data" not found Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.492370 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-operator-scripts\") pod \"neutron2d2f-account-delete-5c2sd\" (UID: \"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee\") " pod="openstack/neutron2d2f-account-delete-5c2sd" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.492455 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement2704-account-delete-ck9t2"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.502104 4799 generic.go:334] "Generic (PLEG): container finished" podID="bbfb60f9-6e71-4b22-9d74-cab607a00c20" containerID="a9ca7d86df080d4b606da7e5a800c32f1929eb3f6ccde8a25eb20a206bdf9fda" exitCode=2 Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.502150 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bbfb60f9-6e71-4b22-9d74-cab607a00c20","Type":"ContainerDied","Data":"a9ca7d86df080d4b606da7e5a800c32f1929eb3f6ccde8a25eb20a206bdf9fda"} Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.502205 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="bbfb60f9-6e71-4b22-9d74-cab607a00c20" containerName="ovsdbserver-nb" containerID="cri-o://7d643d949a4bb973305b6b986dbd4165fc4a1f0f0846279bc0b5ddf8929ad32b" gracePeriod=300 Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.524162 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zn9g\" (UniqueName: \"kubernetes.io/projected/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-kube-api-access-8zn9g\") pod \"neutron2d2f-account-delete-5c2sd\" (UID: \"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee\") " pod="openstack/neutron2d2f-account-delete-5c2sd" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.577875 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell000c8-account-delete-n5mhs"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.579753 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell000c8-account-delete-n5mhs" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.600516 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5b78\" (UniqueName: \"kubernetes.io/projected/87439e47-6c84-4467-a49b-03ed6e86a9eb-kube-api-access-v5b78\") pod \"novacell000c8-account-delete-n5mhs\" (UID: \"87439e47-6c84-4467-a49b-03ed6e86a9eb\") " pod="openstack/novacell000c8-account-delete-n5mhs" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.600921 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts\") pod \"novacell000c8-account-delete-n5mhs\" (UID: \"87439e47-6c84-4467-a49b-03ed6e86a9eb\") " pod="openstack/novacell000c8-account-delete-n5mhs" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.600995 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nkk7\" (UniqueName: \"kubernetes.io/projected/b6f22313-6d9a-4b35-b892-695903d354b1-kube-api-access-8nkk7\") pod \"placement2704-account-delete-ck9t2\" (UID: \"b6f22313-6d9a-4b35-b892-695903d354b1\") " pod="openstack/placement2704-account-delete-ck9t2" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.601155 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts\") pod \"placement2704-account-delete-ck9t2\" (UID: \"b6f22313-6d9a-4b35-b892-695903d354b1\") " pod="openstack/placement2704-account-delete-ck9t2" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.601233 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4cdf\" (UniqueName: \"kubernetes.io/projected/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-kube-api-access-c4cdf\") pod \"barbican44f2-account-delete-2hm9q\" (UID: \"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a\") " pod="openstack/barbican44f2-account-delete-2hm9q" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.601284 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts\") pod \"barbican44f2-account-delete-2hm9q\" (UID: \"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a\") " pod="openstack/barbican44f2-account-delete-2hm9q" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.602156 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts\") pod \"barbican44f2-account-delete-2hm9q\" (UID: \"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a\") " pod="openstack/barbican44f2-account-delete-2hm9q" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.614584 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell000c8-account-delete-n5mhs"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.622172 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts\") pod \"placement2704-account-delete-ck9t2\" (UID: \"b6f22313-6d9a-4b35-b892-695903d354b1\") " pod="openstack/placement2704-account-delete-ck9t2" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.633330 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4cdf\" (UniqueName: \"kubernetes.io/projected/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-kube-api-access-c4cdf\") pod \"barbican44f2-account-delete-2hm9q\" (UID: \"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a\") " pod="openstack/barbican44f2-account-delete-2hm9q" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.664857 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-kjxxs"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.670966 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nkk7\" (UniqueName: \"kubernetes.io/projected/b6f22313-6d9a-4b35-b892-695903d354b1-kube-api-access-8nkk7\") pod \"placement2704-account-delete-ck9t2\" (UID: \"b6f22313-6d9a-4b35-b892-695903d354b1\") " pod="openstack/placement2704-account-delete-ck9t2" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.694111 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-kjxxs"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.707532 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5b78\" (UniqueName: \"kubernetes.io/projected/87439e47-6c84-4467-a49b-03ed6e86a9eb-kube-api-access-v5b78\") pod \"novacell000c8-account-delete-n5mhs\" (UID: \"87439e47-6c84-4467-a49b-03ed6e86a9eb\") " pod="openstack/novacell000c8-account-delete-n5mhs" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.707569 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts\") pod \"novacell000c8-account-delete-n5mhs\" (UID: \"87439e47-6c84-4467-a49b-03ed6e86a9eb\") " pod="openstack/novacell000c8-account-delete-n5mhs" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.708940 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts\") pod \"novacell000c8-account-delete-n5mhs\" (UID: \"87439e47-6c84-4467-a49b-03ed6e86a9eb\") " pod="openstack/novacell000c8-account-delete-n5mhs" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.742775 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell125d6-account-delete-cw4x8"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.744083 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell125d6-account-delete-cw4x8" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.745612 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5b78\" (UniqueName: \"kubernetes.io/projected/87439e47-6c84-4467-a49b-03ed6e86a9eb-kube-api-access-v5b78\") pod \"novacell000c8-account-delete-n5mhs\" (UID: \"87439e47-6c84-4467-a49b-03ed6e86a9eb\") " pod="openstack/novacell000c8-account-delete-n5mhs" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.765232 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron2d2f-account-delete-5c2sd" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.777924 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell125d6-account-delete-cw4x8"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.803773 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican44f2-account-delete-2hm9q" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.807872 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi722e-account-delete-th6f8"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.808873 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4a27166-5918-4ee7-8569-a9f58914d7be-operator-scripts\") pod \"novacell125d6-account-delete-cw4x8\" (UID: \"d4a27166-5918-4ee7-8569-a9f58914d7be\") " pod="openstack/novacell125d6-account-delete-cw4x8" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.808945 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvfj6\" (UniqueName: \"kubernetes.io/projected/d4a27166-5918-4ee7-8569-a9f58914d7be-kube-api-access-fvfj6\") pod \"novacell125d6-account-delete-cw4x8\" (UID: \"d4a27166-5918-4ee7-8569-a9f58914d7be\") " pod="openstack/novacell125d6-account-delete-cw4x8" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.809154 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi722e-account-delete-th6f8" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.816866 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-fxr57"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.817068 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-fxr57" podUID="09082d4e-0623-40c5-b584-1eb58985cfc0" containerName="openstack-network-exporter" containerID="cri-o://145fcceaa41ce0287f7a6fe79891e468626f0454f5b1fca70bfabb879bb1c5fa" gracePeriod=30 Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.827800 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi722e-account-delete-th6f8"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.832918 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement2704-account-delete-ck9t2" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.839458 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-2dz2n"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.888231 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-kgrrs"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.912632 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5sfx\" (UniqueName: \"kubernetes.io/projected/63757cd4-2ed0-4423-b466-4dfac76e3ac4-kube-api-access-j5sfx\") pod \"novaapi722e-account-delete-th6f8\" (UID: \"63757cd4-2ed0-4423-b466-4dfac76e3ac4\") " pod="openstack/novaapi722e-account-delete-th6f8" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.912714 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4a27166-5918-4ee7-8569-a9f58914d7be-operator-scripts\") pod \"novacell125d6-account-delete-cw4x8\" (UID: \"d4a27166-5918-4ee7-8569-a9f58914d7be\") " pod="openstack/novacell125d6-account-delete-cw4x8" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.912766 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvfj6\" (UniqueName: \"kubernetes.io/projected/d4a27166-5918-4ee7-8569-a9f58914d7be-kube-api-access-fvfj6\") pod \"novacell125d6-account-delete-cw4x8\" (UID: \"d4a27166-5918-4ee7-8569-a9f58914d7be\") " pod="openstack/novacell125d6-account-delete-cw4x8" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.912827 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63757cd4-2ed0-4423-b466-4dfac76e3ac4-operator-scripts\") pod \"novaapi722e-account-delete-th6f8\" (UID: \"63757cd4-2ed0-4423-b466-4dfac76e3ac4\") " pod="openstack/novaapi722e-account-delete-th6f8" Nov 24 07:09:10 crc kubenswrapper[4799]: E1124 07:09:10.912982 4799 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 07:09:10 crc kubenswrapper[4799]: E1124 07:09:10.913018 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data podName:366b7548-2fca-4623-a45c-c1e0367ce93a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:11.913006107 +0000 UTC m=+1297.568988581 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data") pod "rabbitmq-cell1-server-0" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a") : configmap "rabbitmq-cell1-config-data" not found Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.913910 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4a27166-5918-4ee7-8569-a9f58914d7be-operator-scripts\") pod \"novacell125d6-account-delete-cw4x8\" (UID: \"d4a27166-5918-4ee7-8569-a9f58914d7be\") " pod="openstack/novacell125d6-account-delete-cw4x8" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.918762 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-gpvs7"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.937660 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-gpvs7"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.947333 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvfj6\" (UniqueName: \"kubernetes.io/projected/d4a27166-5918-4ee7-8569-a9f58914d7be-kube-api-access-fvfj6\") pod \"novacell125d6-account-delete-cw4x8\" (UID: \"d4a27166-5918-4ee7-8569-a9f58914d7be\") " pod="openstack/novacell125d6-account-delete-cw4x8" Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.991928 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 07:09:10 crc kubenswrapper[4799]: I1124 07:09:10.992587 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" containerName="openstack-network-exporter" containerID="cri-o://2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c" gracePeriod=300 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.002130 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.002397 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cb94fabc-dfc3-40cc-9791-0272924692da" containerName="cinder-scheduler" containerID="cri-o://dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.002800 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cb94fabc-dfc3-40cc-9791-0272924692da" containerName="probe" containerID="cri-o://af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.010084 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell000c8-account-delete-n5mhs" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.010654 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-7mn2f"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.010821 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" podUID="ce1d3095-b344-4e6c-b49a-d1de3a7260ee" containerName="dnsmasq-dns" containerID="cri-o://38b78508876ab5b361ddd58cb9e13822b959fef1c0bad5d5fe801de1bd871409" gracePeriod=10 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.015614 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63757cd4-2ed0-4423-b466-4dfac76e3ac4-operator-scripts\") pod \"novaapi722e-account-delete-th6f8\" (UID: \"63757cd4-2ed0-4423-b466-4dfac76e3ac4\") " pod="openstack/novaapi722e-account-delete-th6f8" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.015697 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5sfx\" (UniqueName: \"kubernetes.io/projected/63757cd4-2ed0-4423-b466-4dfac76e3ac4-kube-api-access-j5sfx\") pod \"novaapi722e-account-delete-th6f8\" (UID: \"63757cd4-2ed0-4423-b466-4dfac76e3ac4\") " pod="openstack/novaapi722e-account-delete-th6f8" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.016834 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63757cd4-2ed0-4423-b466-4dfac76e3ac4-operator-scripts\") pod \"novaapi722e-account-delete-th6f8\" (UID: \"63757cd4-2ed0-4423-b466-4dfac76e3ac4\") " pod="openstack/novaapi722e-account-delete-th6f8" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.026924 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-hj8ls"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.040923 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-hj8ls"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.049387 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.049628 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f643414c-72f0-4e7c-b804-20a56f3207ab" containerName="cinder-api-log" containerID="cri-o://1bd8daa785503f4e2fcb67f8d5e0fa713f93ffd5b803f47263907f5420e39c50" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.050002 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f643414c-72f0-4e7c-b804-20a56f3207ab" containerName="cinder-api" containerID="cri-o://30959909b72adeb936fc238d713e097048621efd99d9e5661c63d6da5128b9e8" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.060511 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5sfx\" (UniqueName: \"kubernetes.io/projected/63757cd4-2ed0-4423-b466-4dfac76e3ac4-kube-api-access-j5sfx\") pod \"novaapi722e-account-delete-th6f8\" (UID: \"63757cd4-2ed0-4423-b466-4dfac76e3ac4\") " pod="openstack/novaapi722e-account-delete-th6f8" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.064916 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-d57xl"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.080507 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell125d6-account-delete-cw4x8" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.114380 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-d57xl"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.176101 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-77fc689cdf-xj7m6"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.176546 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-77fc689cdf-xj7m6" podUID="0a50ddac-d55d-4e99-8dea-8519b8d732d6" containerName="neutron-api" containerID="cri-o://02ffa0bb3d383e7c692b859a1e3376a9f3f7187ae40a809812908dbb07a5603c" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.180074 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-77fc689cdf-xj7m6" podUID="0a50ddac-d55d-4e99-8dea-8519b8d732d6" containerName="neutron-httpd" containerID="cri-o://d3a425e1e004a0c21cb397f3983c5fe55824d01453fd599218ea20167f5095a8" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.181103 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi722e-account-delete-th6f8" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.206072 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" containerName="ovsdbserver-sb" containerID="cri-o://477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742" gracePeriod=300 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.210902 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.211211 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" containerName="glance-log" containerID="cri-o://668ba3f136f6d97ba947f42f79e2823ce4845dbde12c05bc8fe71572551507f9" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.211652 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" containerName="glance-httpd" containerID="cri-o://183c3c900b89e4d354f72e7ee6593a3ca4a44de57d248b7f61ac4959fe51c8ef" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.243886 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.244332 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-server" containerID="cri-o://dec54396850ed098f8d988491f08fcf315975d08d63a8517e087677c141683a7" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.244566 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-updater" containerID="cri-o://479a19c5d4397919f77e002bce98bc8cfe8b88d861166ad64dbe89337e0e2126" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.244708 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-auditor" containerID="cri-o://6d52bb110c0d262b0946e49bcf01652ba0241c80b9cad0add83df423d3294499" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.244766 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-reaper" containerID="cri-o://de17e7e1d402f63433a3438f6c8d6c44f1b24885c6c84bac90d88ef38ef8f5f7" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.244896 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-replicator" containerID="cri-o://22c9751adfda0a99f240920e420a888061186c08a987e917be24c8bc2d4dc9d3" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.244949 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-server" containerID="cri-o://8ac4f0ccf8167b086abf61aae9df350c81a1408a802be79619c6747aa2cc47af" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.245013 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-server" containerID="cri-o://5e46da699973a5e22e83b9aaac7e0e2ff7525ee80f32ff63ebc1fc46a7e36544" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.245071 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-updater" containerID="cri-o://c8e47fdf47fa1c6b2fdcb1d77c84b445e782cfd00343ab8b1883a72bed9feebd" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.245108 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-auditor" containerID="cri-o://9b90579a000b03e6467dabd651cbce2abb14ef949a00cfdd47d7596c85645587" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.245140 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-replicator" containerID="cri-o://c01da2c24a6c25ef981a4bf81e1ef3766e00137d2cf3e37995a2696e9e28633c" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.245181 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="rsync" containerID="cri-o://815111e7fbfc05e4da9663fc01168d0881dc15d3f9b8fe7ec1ce5a3f38cc6d40" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.245215 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="swift-recon-cron" containerID="cri-o://57ac5adb075544c22ae8aa32fb5357e1f72bc9a58d99ba17c8bb148b66dade15" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.245252 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-expirer" containerID="cri-o://977a35126c8112db8b3dddf63c96480f5494a7696b5982d214d42310a46c7b83" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.245287 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-replicator" containerID="cri-o://3328a1af0097e404c16240f7713c0a9b7963f642d250b69998ef457311035c34" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.245322 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-auditor" containerID="cri-o://a412226a89e3fd5d662ed68efbfbea22971efde4e4219cd6b56cf841e130fc92" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.273576 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-x9mds"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.281182 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-x9mds"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.301326 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-fj4rj"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.313467 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-fj4rj"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.391591 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.403301 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e9a036d8-ce56-4654-a782-15cd18730378" containerName="glance-log" containerID="cri-o://bb47a5f8acf503e4173095e108c8eaae5b42159d81d170336e69837e3fecd4b9" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.403750 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e9a036d8-ce56-4654-a782-15cd18730378" containerName="glance-httpd" containerID="cri-o://e97f64b93e8c87fb3e9989776d9be9cfbdd54d15ade6362a154f53220314a180" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: E1124 07:09:11.546668 4799 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 07:09:11 crc kubenswrapper[4799]: E1124 07:09:11.546740 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data podName:96acf88b-3fe8-46ac-8393-37f720632dd6 nodeName:}" failed. No retries permitted until 2025-11-24 07:09:13.54672277 +0000 UTC m=+1299.202705244 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data") pod "rabbitmq-server-0" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6") : configmap "rabbitmq-config-data" not found Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.570025 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.580715 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-65f6685878-7bjdh"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.585187 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-65f6685878-7bjdh" podUID="897d0d90-bf27-4d5b-b136-863379656f34" containerName="placement-log" containerID="cri-o://702c77d5a08ac68b078c369149870d0f8a51352018417fd5ffd7488783fd889b" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.585790 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-65f6685878-7bjdh" podUID="897d0d90-bf27-4d5b-b136-863379656f34" containerName="placement-api" containerID="cri-o://62f862c6ff3823a347d374cd9375930cec2ce5b254235146abb2a0f3a2a1a61c" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.643114 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-9445d5865-fhgvf"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.643337 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-9445d5865-fhgvf" podUID="b639ad0e-8438-4a36-a776-ee53b6f0c080" containerName="proxy-httpd" containerID="cri-o://2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.643705 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-9445d5865-fhgvf" podUID="b639ad0e-8438-4a36-a776-ee53b6f0c080" containerName="proxy-server" containerID="cri-o://f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.644371 4799 generic.go:334] "Generic (PLEG): container finished" podID="f643414c-72f0-4e7c-b804-20a56f3207ab" containerID="1bd8daa785503f4e2fcb67f8d5e0fa713f93ffd5b803f47263907f5420e39c50" exitCode=143 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.644413 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f643414c-72f0-4e7c-b804-20a56f3207ab","Type":"ContainerDied","Data":"1bd8daa785503f4e2fcb67f8d5e0fa713f93ffd5b803f47263907f5420e39c50"} Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.741630 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03546c3d-2da5-4476-8c22-5a14939cfd80" path="/var/lib/kubelet/pods/03546c3d-2da5-4476-8c22-5a14939cfd80/volumes" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.756359 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ccb00cb-56e5-4588-8c60-88b52e92721e" path="/var/lib/kubelet/pods/0ccb00cb-56e5-4588-8c60-88b52e92721e/volumes" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.757331 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57aab8d2-548c-47cf-9beb-bbe3958ebce6" path="/var/lib/kubelet/pods/57aab8d2-548c-47cf-9beb-bbe3958ebce6/volumes" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.758016 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="959e83f3-69f2-433b-a326-242ec1c4b226" path="/var/lib/kubelet/pods/959e83f3-69f2-433b-a326-242ec1c4b226/volumes" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.773020 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98cd3086-ad5c-4ca7-bcf7-e68e52ef8053" path="/var/lib/kubelet/pods/98cd3086-ad5c-4ca7-bcf7-e68e52ef8053/volumes" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.774469 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a987addd-6874-4499-9ea1-27ec674bf28a" path="/var/lib/kubelet/pods/a987addd-6874-4499-9ea1-27ec674bf28a/volumes" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.775687 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3cd7765-7004-4a36-8331-029e35639568" path="/var/lib/kubelet/pods/b3cd7765-7004-4a36-8331-029e35639568/volumes" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.776645 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d" path="/var/lib/kubelet/pods/d4cd0282-fb2c-48bf-91cb-ae8bd6e17c6d/volumes" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.777993 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.778066 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.778092 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.778110 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.780801 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-log" containerID="cri-o://c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.781066 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerName="nova-api-log" containerID="cri-o://cbf304821538247ce0d3cfcbea062c7ba81f6603d05b995bc8ebc90500583b62" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.781740 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-metadata" containerID="cri-o://5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.781864 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerName="nova-api-api" containerID="cri-o://f6b04750444de5a3182982a1b53fc81a035e9a9f2eb6df3d21f831648f26fc49" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.781990 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell125d6-account-delete-cw4x8"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.794562 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-25d6-account-create-r6lpd"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.798199 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="479a19c5d4397919f77e002bce98bc8cfe8b88d861166ad64dbe89337e0e2126" exitCode=0 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.798233 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="3328a1af0097e404c16240f7713c0a9b7963f642d250b69998ef457311035c34" exitCode=0 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.798296 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"479a19c5d4397919f77e002bce98bc8cfe8b88d861166ad64dbe89337e0e2126"} Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.798327 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"3328a1af0097e404c16240f7713c0a9b7963f642d250b69998ef457311035c34"} Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.811627 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace/ovsdbserver-sb/0.log" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.811673 4799 generic.go:334] "Generic (PLEG): container finished" podID="2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" containerID="2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c" exitCode=2 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.811753 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace","Type":"ContainerDied","Data":"2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c"} Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.819935 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-25d6-account-create-r6lpd"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.831081 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-mkgcm"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.836985 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-mkgcm"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.852262 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6fbb66bcc8-62525"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.852722 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" podUID="c986e5b7-9864-46ae-9c99-9072e9c6ce34" containerName="barbican-keystone-listener-log" containerID="cri-o://d87462cbe65628447244d9ea1c17fb1adb8cfc23702f9f99482cba9b90403d7c" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.853136 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" podUID="c986e5b7-9864-46ae-9c99-9072e9c6ce34" containerName="barbican-keystone-listener" containerID="cri-o://90b897d4e24c66a97b38d28448dae228ba7620bc06b2cd553b7c3edcb5ea7710" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.884679 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7dd777b755-wgf6d"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.884920 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7dd777b755-wgf6d" podUID="a9ee9f9f-66be-467a-a982-9ede1e1b682c" containerName="barbican-worker-log" containerID="cri-o://51856ae48de556e52c655ad669ed192facc4a302f9d9889f4ff2d00441ecb8fa" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.885279 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7dd777b755-wgf6d" podUID="a9ee9f9f-66be-467a-a982-9ede1e1b682c" containerName="barbican-worker" containerID="cri-o://f9a800a0af5f1a9b50d8af38a76b70c4541840dcf2598a2012116de5e32ec921" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.900201 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bbfb60f9-6e71-4b22-9d74-cab607a00c20/ovsdbserver-nb/0.log" Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.900238 4799 generic.go:334] "Generic (PLEG): container finished" podID="bbfb60f9-6e71-4b22-9d74-cab607a00c20" containerID="7d643d949a4bb973305b6b986dbd4165fc4a1f0f0846279bc0b5ddf8929ad32b" exitCode=143 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.900288 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bbfb60f9-6e71-4b22-9d74-cab607a00c20","Type":"ContainerDied","Data":"7d643d949a4bb973305b6b986dbd4165fc4a1f0f0846279bc0b5ddf8929ad32b"} Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.900470 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="366b7548-2fca-4623-a45c-c1e0367ce93a" containerName="rabbitmq" containerID="cri-o://54abd42b7590aea7f2db6c595fca4f645386857fcbdb38711dfb864deb083510" gracePeriod=604800 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.901987 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-644b86779b-b4psv"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.902586 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-644b86779b-b4psv" podUID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerName="barbican-api-log" containerID="cri-o://c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.909462 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-644b86779b-b4psv" podUID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerName="barbican-api" containerID="cri-o://2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.930648 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.933348 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3feaa8e4-537c-44d2-ae92-ba999eb1626c" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://330d7990c1e4ba99f0cea6c8c2d506420220e3481e87800ebda2781888dd174a" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.943440 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.943715 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0d55de08-d6b7-44f6-8e96-44fcd8999000" containerName="nova-scheduler-scheduler" containerID="cri-o://0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac" gracePeriod=30 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.957004 4799 generic.go:334] "Generic (PLEG): container finished" podID="ce1d3095-b344-4e6c-b49a-d1de3a7260ee" containerID="38b78508876ab5b361ddd58cb9e13822b959fef1c0bad5d5fe801de1bd871409" exitCode=0 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.963787 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="96acf88b-3fe8-46ac-8393-37f720632dd6" containerName="rabbitmq" containerID="cri-o://2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca" gracePeriod=604800 Nov 24 07:09:11 crc kubenswrapper[4799]: I1124 07:09:11.963948 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" event={"ID":"ce1d3095-b344-4e6c-b49a-d1de3a7260ee","Type":"ContainerDied","Data":"38b78508876ab5b361ddd58cb9e13822b959fef1c0bad5d5fe801de1bd871409"} Nov 24 07:09:11 crc kubenswrapper[4799]: E1124 07:09:11.978527 4799 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 07:09:11 crc kubenswrapper[4799]: E1124 07:09:11.978599 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data podName:366b7548-2fca-4623-a45c-c1e0367ce93a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:13.978580395 +0000 UTC m=+1299.634562869 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data") pod "rabbitmq-cell1-server-0" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a") : configmap "rabbitmq-cell1-config-data" not found Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.001249 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-fxr57_09082d4e-0623-40c5-b584-1eb58985cfc0/openstack-network-exporter/0.log" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.001288 4799 generic.go:334] "Generic (PLEG): container finished" podID="09082d4e-0623-40c5-b584-1eb58985cfc0" containerID="145fcceaa41ce0287f7a6fe79891e468626f0454f5b1fca70bfabb879bb1c5fa" exitCode=2 Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.001318 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fxr57" event={"ID":"09082d4e-0623-40c5-b584-1eb58985cfc0","Type":"ContainerDied","Data":"145fcceaa41ce0287f7a6fe79891e468626f0454f5b1fca70bfabb879bb1c5fa"} Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.243257 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bbfb60f9-6e71-4b22-9d74-cab607a00c20/ovsdbserver-nb/0.log" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.243912 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 07:09:12 crc kubenswrapper[4799]: E1124 07:09:12.308022 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.373932 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance00d0-account-delete-kbtpl"] Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.391434 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-combined-ca-bundle\") pod \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.391612 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drg4w\" (UniqueName: \"kubernetes.io/projected/bbfb60f9-6e71-4b22-9d74-cab607a00c20-kube-api-access-drg4w\") pod \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.391643 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.392518 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-metrics-certs-tls-certs\") pod \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.392591 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdbserver-nb-tls-certs\") pod \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.392636 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-scripts\") pod \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.392686 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-config\") pod \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.392703 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdb-rundir\") pod \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\" (UID: \"bbfb60f9-6e71-4b22-9d74-cab607a00c20\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.393480 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "bbfb60f9-6e71-4b22-9d74-cab607a00c20" (UID: "bbfb60f9-6e71-4b22-9d74-cab607a00c20"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.395468 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-config" (OuterVolumeSpecName: "config") pod "bbfb60f9-6e71-4b22-9d74-cab607a00c20" (UID: "bbfb60f9-6e71-4b22-9d74-cab607a00c20"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.396106 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-scripts" (OuterVolumeSpecName: "scripts") pod "bbfb60f9-6e71-4b22-9d74-cab607a00c20" (UID: "bbfb60f9-6e71-4b22-9d74-cab607a00c20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.405159 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbfb60f9-6e71-4b22-9d74-cab607a00c20-kube-api-access-drg4w" (OuterVolumeSpecName: "kube-api-access-drg4w") pod "bbfb60f9-6e71-4b22-9d74-cab607a00c20" (UID: "bbfb60f9-6e71-4b22-9d74-cab607a00c20"). InnerVolumeSpecName "kube-api-access-drg4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.411604 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinderbbb1-account-delete-tjfp2"] Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.412071 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "bbfb60f9-6e71-4b22-9d74-cab607a00c20" (UID: "bbfb60f9-6e71-4b22-9d74-cab607a00c20"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: E1124 07:09:12.436746 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 07:09:12 crc kubenswrapper[4799]: E1124 07:09:12.447334 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 07:09:12 crc kubenswrapper[4799]: E1124 07:09:12.447582 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerName="ovn-northd" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.465667 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbfb60f9-6e71-4b22-9d74-cab607a00c20" (UID: "bbfb60f9-6e71-4b22-9d74-cab607a00c20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.499506 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.499538 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbfb60f9-6e71-4b22-9d74-cab607a00c20-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.499548 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.499557 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.499566 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drg4w\" (UniqueName: \"kubernetes.io/projected/bbfb60f9-6e71-4b22-9d74-cab607a00c20-kube-api-access-drg4w\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.499588 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.527303 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-fxr57_09082d4e-0623-40c5-b584-1eb58985cfc0/openstack-network-exporter/0.log" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.527367 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.537783 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="844d53db-a457-4153-bdb4-d557ba34c099" containerName="galera" containerID="cri-o://57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e" gracePeriod=30 Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.540428 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "bbfb60f9-6e71-4b22-9d74-cab607a00c20" (UID: "bbfb60f9-6e71-4b22-9d74-cab607a00c20"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.601493 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovn-rundir\") pod \"09082d4e-0623-40c5-b584-1eb58985cfc0\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.601592 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-combined-ca-bundle\") pod \"09082d4e-0623-40c5-b584-1eb58985cfc0\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.601629 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09082d4e-0623-40c5-b584-1eb58985cfc0-config\") pod \"09082d4e-0623-40c5-b584-1eb58985cfc0\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.601671 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovs-rundir\") pod \"09082d4e-0623-40c5-b584-1eb58985cfc0\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.601732 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-metrics-certs-tls-certs\") pod \"09082d4e-0623-40c5-b584-1eb58985cfc0\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.601763 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flvrz\" (UniqueName: \"kubernetes.io/projected/09082d4e-0623-40c5-b584-1eb58985cfc0-kube-api-access-flvrz\") pod \"09082d4e-0623-40c5-b584-1eb58985cfc0\" (UID: \"09082d4e-0623-40c5-b584-1eb58985cfc0\") " Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.602183 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.605128 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "09082d4e-0623-40c5-b584-1eb58985cfc0" (UID: "09082d4e-0623-40c5-b584-1eb58985cfc0"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.605433 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "09082d4e-0623-40c5-b584-1eb58985cfc0" (UID: "09082d4e-0623-40c5-b584-1eb58985cfc0"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.607132 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09082d4e-0623-40c5-b584-1eb58985cfc0-config" (OuterVolumeSpecName: "config") pod "09082d4e-0623-40c5-b584-1eb58985cfc0" (UID: "09082d4e-0623-40c5-b584-1eb58985cfc0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.633215 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09082d4e-0623-40c5-b584-1eb58985cfc0-kube-api-access-flvrz" (OuterVolumeSpecName: "kube-api-access-flvrz") pod "09082d4e-0623-40c5-b584-1eb58985cfc0" (UID: "09082d4e-0623-40c5-b584-1eb58985cfc0"). InnerVolumeSpecName "kube-api-access-flvrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.663640 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.685147 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovs-vswitchd" containerID="cri-o://16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" gracePeriod=29 Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.741558 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flvrz\" (UniqueName: \"kubernetes.io/projected/09082d4e-0623-40c5-b584-1eb58985cfc0-kube-api-access-flvrz\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.741584 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.741593 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.741603 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09082d4e-0623-40c5-b584-1eb58985cfc0-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.741627 4799 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/09082d4e-0623-40c5-b584-1eb58985cfc0-ovs-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: E1124 07:09:12.755317 4799 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 24 07:09:12 crc kubenswrapper[4799]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 24 07:09:12 crc kubenswrapper[4799]: + source /usr/local/bin/container-scripts/functions Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNBridge=br-int Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNRemote=tcp:localhost:6642 Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNEncapType=geneve Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNAvailabilityZones= Nov 24 07:09:12 crc kubenswrapper[4799]: ++ EnableChassisAsGateway=true Nov 24 07:09:12 crc kubenswrapper[4799]: ++ PhysicalNetworks= Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNHostName= Nov 24 07:09:12 crc kubenswrapper[4799]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 24 07:09:12 crc kubenswrapper[4799]: ++ ovs_dir=/var/lib/openvswitch Nov 24 07:09:12 crc kubenswrapper[4799]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 24 07:09:12 crc kubenswrapper[4799]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 24 07:09:12 crc kubenswrapper[4799]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + sleep 0.5 Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + sleep 0.5 Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + sleep 0.5 Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + cleanup_ovsdb_server_semaphore Nov 24 07:09:12 crc kubenswrapper[4799]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 07:09:12 crc kubenswrapper[4799]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 24 07:09:12 crc kubenswrapper[4799]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-kgrrs" message=< Nov 24 07:09:12 crc kubenswrapper[4799]: Exiting ovsdb-server (5) [ OK ] Nov 24 07:09:12 crc kubenswrapper[4799]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 24 07:09:12 crc kubenswrapper[4799]: + source /usr/local/bin/container-scripts/functions Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNBridge=br-int Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNRemote=tcp:localhost:6642 Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNEncapType=geneve Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNAvailabilityZones= Nov 24 07:09:12 crc kubenswrapper[4799]: ++ EnableChassisAsGateway=true Nov 24 07:09:12 crc kubenswrapper[4799]: ++ PhysicalNetworks= Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNHostName= Nov 24 07:09:12 crc kubenswrapper[4799]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 24 07:09:12 crc kubenswrapper[4799]: ++ ovs_dir=/var/lib/openvswitch Nov 24 07:09:12 crc kubenswrapper[4799]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 24 07:09:12 crc kubenswrapper[4799]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 24 07:09:12 crc kubenswrapper[4799]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + sleep 0.5 Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + sleep 0.5 Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + sleep 0.5 Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + cleanup_ovsdb_server_semaphore Nov 24 07:09:12 crc kubenswrapper[4799]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 07:09:12 crc kubenswrapper[4799]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 24 07:09:12 crc kubenswrapper[4799]: > Nov 24 07:09:12 crc kubenswrapper[4799]: E1124 07:09:12.755378 4799 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 24 07:09:12 crc kubenswrapper[4799]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 24 07:09:12 crc kubenswrapper[4799]: + source /usr/local/bin/container-scripts/functions Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNBridge=br-int Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNRemote=tcp:localhost:6642 Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNEncapType=geneve Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNAvailabilityZones= Nov 24 07:09:12 crc kubenswrapper[4799]: ++ EnableChassisAsGateway=true Nov 24 07:09:12 crc kubenswrapper[4799]: ++ PhysicalNetworks= Nov 24 07:09:12 crc kubenswrapper[4799]: ++ OVNHostName= Nov 24 07:09:12 crc kubenswrapper[4799]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 24 07:09:12 crc kubenswrapper[4799]: ++ ovs_dir=/var/lib/openvswitch Nov 24 07:09:12 crc kubenswrapper[4799]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 24 07:09:12 crc kubenswrapper[4799]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 24 07:09:12 crc kubenswrapper[4799]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + sleep 0.5 Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + sleep 0.5 Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + sleep 0.5 Nov 24 07:09:12 crc kubenswrapper[4799]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 07:09:12 crc kubenswrapper[4799]: + cleanup_ovsdb_server_semaphore Nov 24 07:09:12 crc kubenswrapper[4799]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 07:09:12 crc kubenswrapper[4799]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 24 07:09:12 crc kubenswrapper[4799]: > pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server" containerID="cri-o://343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.755454 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server" containerID="cri-o://343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" gracePeriod=29 Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.829066 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "bbfb60f9-6e71-4b22-9d74-cab607a00c20" (UID: "bbfb60f9-6e71-4b22-9d74-cab607a00c20"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.846182 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbfb60f9-6e71-4b22-9d74-cab607a00c20-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.938071 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.939373 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.941029 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09082d4e-0623-40c5-b584-1eb58985cfc0" (UID: "09082d4e-0623-40c5-b584-1eb58985cfc0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.942515 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-k9t68"] Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.948049 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.954458 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-k9t68"] Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.964162 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.965975 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="168550d3-6597-41ff-adf4-d20f190c7a02" containerName="nova-cell1-conductor-conductor" containerID="cri-o://192c554265f38b0137575a7bb2ad8ef843ac317c21ba65066896ec0465f061f3" gracePeriod=30 Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.972800 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace/ovsdbserver-sb/0.log" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.973094 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 07:09:12 crc kubenswrapper[4799]: I1124 07:09:12.992555 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8m4cw"] Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.012531 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8m4cw"] Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.027703 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.028339 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="66ae5dfe-6c96-4d2d-99cf-b792b139d1a1" containerName="nova-cell0-conductor-conductor" containerID="cri-o://8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05" gracePeriod=30 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.046986 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055590 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config-secret\") pod \"76771b43-063e-4bc4-a366-3a13f27e3124\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055652 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-swift-storage-0\") pod \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055675 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-metrics-certs-tls-certs\") pod \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055739 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055775 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-config\") pod \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055793 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xf759\" (UniqueName: \"kubernetes.io/projected/76771b43-063e-4bc4-a366-3a13f27e3124-kube-api-access-xf759\") pod \"76771b43-063e-4bc4-a366-3a13f27e3124\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055817 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-config\") pod \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055840 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-sb\") pod \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055899 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdbserver-sb-tls-certs\") pod \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055951 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79m49\" (UniqueName: \"kubernetes.io/projected/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-kube-api-access-79m49\") pod \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055966 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq46q\" (UniqueName: \"kubernetes.io/projected/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-kube-api-access-mq46q\") pod \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.055992 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-scripts\") pod \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.056042 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-combined-ca-bundle\") pod \"76771b43-063e-4bc4-a366-3a13f27e3124\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.056067 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-svc\") pod \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.056109 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-nb\") pod \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\" (UID: \"ce1d3095-b344-4e6c-b49a-d1de3a7260ee\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.056138 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config\") pod \"76771b43-063e-4bc4-a366-3a13f27e3124\" (UID: \"76771b43-063e-4bc4-a366-3a13f27e3124\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.056188 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdb-rundir\") pod \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.056217 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-combined-ca-bundle\") pod \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\" (UID: \"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.067665 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-config" (OuterVolumeSpecName: "config") pod "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" (UID: "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.076319 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-scripts" (OuterVolumeSpecName: "scripts") pod "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" (UID: "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.076730 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" (UID: "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.087958 4799 generic.go:334] "Generic (PLEG): container finished" podID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerID="c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af" exitCode=143 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.088144 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644b86779b-b4psv" event={"ID":"ebfe77df-bbcd-412f-9703-78e23d93f410","Type":"ContainerDied","Data":"c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.090619 4799 generic.go:334] "Generic (PLEG): container finished" podID="e9a036d8-ce56-4654-a782-15cd18730378" containerID="bb47a5f8acf503e4173095e108c8eaae5b42159d81d170336e69837e3fecd4b9" exitCode=143 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.090675 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e9a036d8-ce56-4654-a782-15cd18730378","Type":"ContainerDied","Data":"bb47a5f8acf503e4173095e108c8eaae5b42159d81d170336e69837e3fecd4b9"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.091226 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" (UID: "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.093696 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance00d0-account-delete-kbtpl" event={"ID":"ad75f933-ab36-4c9d-857d-c2cc11702dd3","Type":"ContainerStarted","Data":"0b036e7485bdbc7a10091975ed6eae68dfe9ad069cb89ee4aca3cd050c565d50"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.093745 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance00d0-account-delete-kbtpl" event={"ID":"ad75f933-ab36-4c9d-857d-c2cc11702dd3","Type":"ContainerStarted","Data":"2ea3adf81d2d683132eb99305e52fcbd10295056e7f7ab9f970e55c50770b682"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.095869 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-kube-api-access-mq46q" (OuterVolumeSpecName: "kube-api-access-mq46q") pod "ce1d3095-b344-4e6c-b49a-d1de3a7260ee" (UID: "ce1d3095-b344-4e6c-b49a-d1de3a7260ee"). InnerVolumeSpecName "kube-api-access-mq46q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.099051 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-kube-api-access-79m49" (OuterVolumeSpecName: "kube-api-access-79m49") pod "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" (UID: "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace"). InnerVolumeSpecName "kube-api-access-79m49". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.101463 4799 generic.go:334] "Generic (PLEG): container finished" podID="c986e5b7-9864-46ae-9c99-9072e9c6ce34" containerID="90b897d4e24c66a97b38d28448dae228ba7620bc06b2cd553b7c3edcb5ea7710" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.101492 4799 generic.go:334] "Generic (PLEG): container finished" podID="c986e5b7-9864-46ae-9c99-9072e9c6ce34" containerID="d87462cbe65628447244d9ea1c17fb1adb8cfc23702f9f99482cba9b90403d7c" exitCode=143 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.101550 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" event={"ID":"c986e5b7-9864-46ae-9c99-9072e9c6ce34","Type":"ContainerDied","Data":"90b897d4e24c66a97b38d28448dae228ba7620bc06b2cd553b7c3edcb5ea7710"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.101577 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" event={"ID":"c986e5b7-9864-46ae-9c99-9072e9c6ce34","Type":"ContainerDied","Data":"d87462cbe65628447244d9ea1c17fb1adb8cfc23702f9f99482cba9b90403d7c"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.102601 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76771b43-063e-4bc4-a366-3a13f27e3124-kube-api-access-xf759" (OuterVolumeSpecName: "kube-api-access-xf759") pod "76771b43-063e-4bc4-a366-3a13f27e3124" (UID: "76771b43-063e-4bc4-a366-3a13f27e3124"). InnerVolumeSpecName "kube-api-access-xf759". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.117447 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance00d0-account-delete-kbtpl" podStartSLOduration=4.117426222 podStartE2EDuration="4.117426222s" podCreationTimestamp="2025-11-24 07:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:09:13.115659062 +0000 UTC m=+1298.771641626" watchObservedRunningTime="2025-11-24 07:09:13.117426222 +0000 UTC m=+1298.773408696" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.126808 4799 generic.go:334] "Generic (PLEG): container finished" podID="cb94fabc-dfc3-40cc-9791-0272924692da" containerID="af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.126899 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb94fabc-dfc3-40cc-9791-0272924692da","Type":"ContainerDied","Data":"af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.130685 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-fxr57_09082d4e-0623-40c5-b584-1eb58985cfc0/openstack-network-exporter/0.log" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.130762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fxr57" event={"ID":"09082d4e-0623-40c5-b584-1eb58985cfc0","Type":"ContainerDied","Data":"cbe00e7f26e6d398423c98a60db1c3f591d35a88157f4ac3ab54a88e3f4aaaf4"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.130800 4799 scope.go:117] "RemoveContainer" containerID="145fcceaa41ce0287f7a6fe79891e468626f0454f5b1fca70bfabb879bb1c5fa" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.130839 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fxr57" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.146118 4799 generic.go:334] "Generic (PLEG): container finished" podID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerID="c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547" exitCode=143 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.146197 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6218522b-3df9-45b0-8edb-fb935a7f4109","Type":"ContainerDied","Data":"c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.149026 4799 generic.go:334] "Generic (PLEG): container finished" podID="897d0d90-bf27-4d5b-b136-863379656f34" containerID="702c77d5a08ac68b078c369149870d0f8a51352018417fd5ffd7488783fd889b" exitCode=143 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.149093 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65f6685878-7bjdh" event={"ID":"897d0d90-bf27-4d5b-b136-863379656f34","Type":"ContainerDied","Data":"702c77d5a08ac68b078c369149870d0f8a51352018417fd5ffd7488783fd889b"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.151713 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace/ovsdbserver-sb/0.log" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.151755 4799 generic.go:334] "Generic (PLEG): container finished" podID="2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" containerID="477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742" exitCode=143 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.151807 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace","Type":"ContainerDied","Data":"477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.151833 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace","Type":"ContainerDied","Data":"eb13270ff2460e3f3e8b1c5606ce701360ed59511f174e9b8db1cdf0a8e86f75"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.151965 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.157120 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bbfb60f9-6e71-4b22-9d74-cab607a00c20/ovsdbserver-nb/0.log" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.157198 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bbfb60f9-6e71-4b22-9d74-cab607a00c20","Type":"ContainerDied","Data":"291ecd9e8f24fbcb763f8e7895bcbae09f1ba8eaa22a90e8c62932f922733e6e"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.157491 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.161886 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-run-httpd\") pod \"b639ad0e-8438-4a36-a776-ee53b6f0c080\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.161947 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-internal-tls-certs\") pod \"b639ad0e-8438-4a36-a776-ee53b6f0c080\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.161975 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-etc-swift\") pod \"b639ad0e-8438-4a36-a776-ee53b6f0c080\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.162004 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-log-httpd\") pod \"b639ad0e-8438-4a36-a776-ee53b6f0c080\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.162047 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dq7j4\" (UniqueName: \"kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-kube-api-access-dq7j4\") pod \"b639ad0e-8438-4a36-a776-ee53b6f0c080\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.162075 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-combined-ca-bundle\") pod \"b639ad0e-8438-4a36-a776-ee53b6f0c080\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.162608 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-config-data\") pod \"b639ad0e-8438-4a36-a776-ee53b6f0c080\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.162830 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-public-tls-certs\") pod \"b639ad0e-8438-4a36-a776-ee53b6f0c080\" (UID: \"b639ad0e-8438-4a36-a776-ee53b6f0c080\") " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.163482 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.163505 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.163561 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xf759\" (UniqueName: \"kubernetes.io/projected/76771b43-063e-4bc4-a366-3a13f27e3124-kube-api-access-xf759\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.163572 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79m49\" (UniqueName: \"kubernetes.io/projected/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-kube-api-access-79m49\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.163581 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq46q\" (UniqueName: \"kubernetes.io/projected/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-kube-api-access-mq46q\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.163589 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.163626 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.164888 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b639ad0e-8438-4a36-a776-ee53b6f0c080" (UID: "b639ad0e-8438-4a36-a776-ee53b6f0c080"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.164955 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b639ad0e-8438-4a36-a776-ee53b6f0c080" (UID: "b639ad0e-8438-4a36-a776-ee53b6f0c080"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.165105 4799 generic.go:334] "Generic (PLEG): container finished" podID="76771b43-063e-4bc4-a366-3a13f27e3124" containerID="1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259" exitCode=137 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.165226 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.169039 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" event={"ID":"ce1d3095-b344-4e6c-b49a-d1de3a7260ee","Type":"ContainerDied","Data":"0f59b71e9d63beedde31ac4aa84625c01cd71aaf89c78e2feb94ca1cbe9291ec"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.169126 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-7mn2f" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.178112 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b639ad0e-8438-4a36-a776-ee53b6f0c080" (UID: "b639ad0e-8438-4a36-a776-ee53b6f0c080"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.179654 4799 generic.go:334] "Generic (PLEG): container finished" podID="a9ee9f9f-66be-467a-a982-9ede1e1b682c" containerID="f9a800a0af5f1a9b50d8af38a76b70c4541840dcf2598a2012116de5e32ec921" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.179686 4799 generic.go:334] "Generic (PLEG): container finished" podID="a9ee9f9f-66be-467a-a982-9ede1e1b682c" containerID="51856ae48de556e52c655ad669ed192facc4a302f9d9889f4ff2d00441ecb8fa" exitCode=143 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.179732 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7dd777b755-wgf6d" event={"ID":"a9ee9f9f-66be-467a-a982-9ede1e1b682c","Type":"ContainerDied","Data":"f9a800a0af5f1a9b50d8af38a76b70c4541840dcf2598a2012116de5e32ec921"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.179785 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7dd777b755-wgf6d" event={"ID":"a9ee9f9f-66be-467a-a982-9ede1e1b682c","Type":"ContainerDied","Data":"51856ae48de556e52c655ad669ed192facc4a302f9d9889f4ff2d00441ecb8fa"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.180216 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-kube-api-access-dq7j4" (OuterVolumeSpecName: "kube-api-access-dq7j4") pod "b639ad0e-8438-4a36-a776-ee53b6f0c080" (UID: "b639ad0e-8438-4a36-a776-ee53b6f0c080"). InnerVolumeSpecName "kube-api-access-dq7j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.185975 4799 generic.go:334] "Generic (PLEG): container finished" podID="b639ad0e-8438-4a36-a776-ee53b6f0c080" containerID="f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.185999 4799 generic.go:334] "Generic (PLEG): container finished" podID="b639ad0e-8438-4a36-a776-ee53b6f0c080" containerID="2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.186199 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-9445d5865-fhgvf" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.186337 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-9445d5865-fhgvf" event={"ID":"b639ad0e-8438-4a36-a776-ee53b6f0c080","Type":"ContainerDied","Data":"f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.186388 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-9445d5865-fhgvf" event={"ID":"b639ad0e-8438-4a36-a776-ee53b6f0c080","Type":"ContainerDied","Data":"2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.186399 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-9445d5865-fhgvf" event={"ID":"b639ad0e-8438-4a36-a776-ee53b6f0c080","Type":"ContainerDied","Data":"b83ebe9c11737bdf41487c483cafea50aac2f9991b0f568474b028a6b4d951a4"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.191425 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderbbb1-account-delete-tjfp2" event={"ID":"d9064d19-6806-4e78-beb4-cda924fe3010","Type":"ContainerStarted","Data":"eff18abe0edf7afb956eb450e236f49b9a829daf1324769ee95e5fc25e5bf28c"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.200920 4799 generic.go:334] "Generic (PLEG): container finished" podID="3feaa8e4-537c-44d2-ae92-ba999eb1626c" containerID="330d7990c1e4ba99f0cea6c8c2d506420220e3481e87800ebda2781888dd174a" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.201025 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3feaa8e4-537c-44d2-ae92-ba999eb1626c","Type":"ContainerDied","Data":"330d7990c1e4ba99f0cea6c8c2d506420220e3481e87800ebda2781888dd174a"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.204674 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.207585 4799 generic.go:334] "Generic (PLEG): container finished" podID="1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" containerID="668ba3f136f6d97ba947f42f79e2823ce4845dbde12c05bc8fe71572551507f9" exitCode=143 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.207635 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf","Type":"ContainerDied","Data":"668ba3f136f6d97ba947f42f79e2823ce4845dbde12c05bc8fe71572551507f9"} Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.255203 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.265242 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.265272 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.265283 4799 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.265291 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b639ad0e-8438-4a36-a776-ee53b6f0c080-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.265299 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dq7j4\" (UniqueName: \"kubernetes.io/projected/b639ad0e-8438-4a36-a776-ee53b6f0c080-kube-api-access-dq7j4\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.269619 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.273614 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.273692 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="0d55de08-d6b7-44f6-8e96-44fcd8999000" containerName="nova-scheduler-scheduler" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.275274 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "76771b43-063e-4bc4-a366-3a13f27e3124" (UID: "76771b43-063e-4bc4-a366-3a13f27e3124"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287097 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="c8e47fdf47fa1c6b2fdcb1d77c84b445e782cfd00343ab8b1883a72bed9feebd" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287122 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="9b90579a000b03e6467dabd651cbce2abb14ef949a00cfdd47d7596c85645587" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287130 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="c01da2c24a6c25ef981a4bf81e1ef3766e00137d2cf3e37995a2696e9e28633c" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287136 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="5e46da699973a5e22e83b9aaac7e0e2ff7525ee80f32ff63ebc1fc46a7e36544" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287142 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="815111e7fbfc05e4da9663fc01168d0881dc15d3f9b8fe7ec1ce5a3f38cc6d40" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287150 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="977a35126c8112db8b3dddf63c96480f5494a7696b5982d214d42310a46c7b83" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287156 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="6d52bb110c0d262b0946e49bcf01652ba0241c80b9cad0add83df423d3294499" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287163 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="22c9751adfda0a99f240920e420a888061186c08a987e917be24c8bc2d4dc9d3" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287169 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="8ac4f0ccf8167b086abf61aae9df350c81a1408a802be79619c6747aa2cc47af" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287175 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="de17e7e1d402f63433a3438f6c8d6c44f1b24885c6c84bac90d88ef38ef8f5f7" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287182 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="a412226a89e3fd5d662ed68efbfbea22971efde4e4219cd6b56cf841e130fc92" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287189 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="dec54396850ed098f8d988491f08fcf315975d08d63a8517e087677c141683a7" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287228 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"c8e47fdf47fa1c6b2fdcb1d77c84b445e782cfd00343ab8b1883a72bed9feebd"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287255 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"9b90579a000b03e6467dabd651cbce2abb14ef949a00cfdd47d7596c85645587"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287265 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"c01da2c24a6c25ef981a4bf81e1ef3766e00137d2cf3e37995a2696e9e28633c"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287273 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"5e46da699973a5e22e83b9aaac7e0e2ff7525ee80f32ff63ebc1fc46a7e36544"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"815111e7fbfc05e4da9663fc01168d0881dc15d3f9b8fe7ec1ce5a3f38cc6d40"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287291 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"977a35126c8112db8b3dddf63c96480f5494a7696b5982d214d42310a46c7b83"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287299 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"6d52bb110c0d262b0946e49bcf01652ba0241c80b9cad0add83df423d3294499"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287307 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"22c9751adfda0a99f240920e420a888061186c08a987e917be24c8bc2d4dc9d3"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287315 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"8ac4f0ccf8167b086abf61aae9df350c81a1408a802be79619c6747aa2cc47af"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287323 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"de17e7e1d402f63433a3438f6c8d6c44f1b24885c6c84bac90d88ef38ef8f5f7"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287333 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"a412226a89e3fd5d662ed68efbfbea22971efde4e4219cd6b56cf841e130fc92"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.287341 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"dec54396850ed098f8d988491f08fcf315975d08d63a8517e087677c141683a7"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.292560 4799 generic.go:334] "Generic (PLEG): container finished" podID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerID="cbf304821538247ce0d3cfcbea062c7ba81f6603d05b995bc8ebc90500583b62" exitCode=143 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.292608 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fe03a76-2eb9-4728-9396-9d2dc1e743af","Type":"ContainerDied","Data":"cbf304821538247ce0d3cfcbea062c7ba81f6603d05b995bc8ebc90500583b62"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.338282 4799 generic.go:334] "Generic (PLEG): container finished" podID="0a50ddac-d55d-4e99-8dea-8519b8d732d6" containerID="d3a425e1e004a0c21cb397f3983c5fe55824d01453fd599218ea20167f5095a8" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.338377 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77fc689cdf-xj7m6" event={"ID":"0a50ddac-d55d-4e99-8dea-8519b8d732d6","Type":"ContainerDied","Data":"d3a425e1e004a0c21cb397f3983c5fe55824d01453fd599218ea20167f5095a8"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.346984 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "09082d4e-0623-40c5-b584-1eb58985cfc0" (UID: "09082d4e-0623-40c5-b584-1eb58985cfc0"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.349826 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" (UID: "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.365166 4799 generic.go:334] "Generic (PLEG): container finished" podID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" exitCode=0 Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.365216 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kgrrs" event={"ID":"a7742c81-45e0-4ef5-aceb-7a631bf9cc37","Type":"ContainerDied","Data":"343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81"} Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.367077 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.367107 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09082d4e-0623-40c5-b584-1eb58985cfc0-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.367118 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.435003 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "76771b43-063e-4bc4-a366-3a13f27e3124" (UID: "76771b43-063e-4bc4-a366-3a13f27e3124"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.457066 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76771b43-063e-4bc4-a366-3a13f27e3124" (UID: "76771b43-063e-4bc4-a366-3a13f27e3124"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.469564 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.469594 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/76771b43-063e-4bc4-a366-3a13f27e3124-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.521232 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b639ad0e-8438-4a36-a776-ee53b6f0c080" (UID: "b639ad0e-8438-4a36-a776-ee53b6f0c080"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.554895 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-config" (OuterVolumeSpecName: "config") pod "ce1d3095-b344-4e6c-b49a-d1de3a7260ee" (UID: "ce1d3095-b344-4e6c-b49a-d1de3a7260ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.571524 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.571790 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.571940 4799 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.572087 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data podName:96acf88b-3fe8-46ac-8393-37f720632dd6 nodeName:}" failed. No retries permitted until 2025-11-24 07:09:17.572066573 +0000 UTC m=+1303.228049047 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data") pod "rabbitmq-server-0" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6") : configmap "rabbitmq-config-data" not found Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.614072 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e is running failed: container process not found" containerID="57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.616658 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e is running failed: container process not found" containerID="57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.622293 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e is running failed: container process not found" containerID="57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.622488 4799 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e is running failed: container process not found" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="844d53db-a457-4153-bdb4-d557ba34c099" containerName="galera" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.647175 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-config-data" (OuterVolumeSpecName: "config-data") pod "b639ad0e-8438-4a36-a776-ee53b6f0c080" (UID: "b639ad0e-8438-4a36-a776-ee53b6f0c080"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.650581 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ce1d3095-b344-4e6c-b49a-d1de3a7260ee" (UID: "ce1d3095-b344-4e6c-b49a-d1de3a7260ee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.667236 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b639ad0e-8438-4a36-a776-ee53b6f0c080" (UID: "b639ad0e-8438-4a36-a776-ee53b6f0c080"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.668679 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" (UID: "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.670910 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cde9df8-eba6-48ec-bc68-ed1fae49df81" path="/var/lib/kubelet/pods/5cde9df8-eba6-48ec-bc68-ed1fae49df81/volumes" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.671670 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="650550d4-96f8-46dd-9117-d4f9c2cdcd9d" path="/var/lib/kubelet/pods/650550d4-96f8-46dd-9117-d4f9c2cdcd9d/volumes" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.672420 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76771b43-063e-4bc4-a366-3a13f27e3124" path="/var/lib/kubelet/pods/76771b43-063e-4bc4-a366-3a13f27e3124/volumes" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.673872 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bea0e41-5c29-42ed-9004-f72e9e7e22cd" path="/var/lib/kubelet/pods/7bea0e41-5c29-42ed-9004-f72e9e7e22cd/volumes" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.674144 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.674216 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.674231 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.674247 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.675062 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e733dc01-8d9d-4a0b-b2e7-e18147562e11" path="/var/lib/kubelet/pods/e733dc01-8d9d-4a0b-b2e7-e18147562e11/volumes" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.676076 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ce1d3095-b344-4e6c-b49a-d1de3a7260ee" (UID: "ce1d3095-b344-4e6c-b49a-d1de3a7260ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.685567 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ce1d3095-b344-4e6c-b49a-d1de3a7260ee" (UID: "ce1d3095-b344-4e6c-b49a-d1de3a7260ee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.686881 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ce1d3095-b344-4e6c-b49a-d1de3a7260ee" (UID: "ce1d3095-b344-4e6c-b49a-d1de3a7260ee"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.740588 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" (UID: "2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.776878 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.776958 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.776973 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.776985 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce1d3095-b344-4e6c-b49a-d1de3a7260ee-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.784440 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b639ad0e-8438-4a36-a776-ee53b6f0c080" (UID: "b639ad0e-8438-4a36-a776-ee53b6f0c080"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.878318 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b639ad0e-8438-4a36-a776-ee53b6f0c080-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.948598 4799 scope.go:117] "RemoveContainer" containerID="2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c" Nov 24 07:09:13 crc kubenswrapper[4799]: I1124 07:09:13.967270 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.983619 4799 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 07:09:13 crc kubenswrapper[4799]: E1124 07:09:13.984276 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data podName:366b7548-2fca-4623-a45c-c1e0367ce93a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:17.98425767 +0000 UTC m=+1303.640240144 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data") pod "rabbitmq-cell1-server-0" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a") : configmap "rabbitmq-cell1-config-data" not found Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.000445 4799 scope.go:117] "RemoveContainer" containerID="477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.004532 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.043652 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.060146 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-7mn2f"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.063316 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.063576 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.073499 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-7mn2f"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.083075 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-fxr57"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085429 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-combined-ca-bundle\") pod \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085473 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8fx9\" (UniqueName: \"kubernetes.io/projected/3feaa8e4-537c-44d2-ae92-ba999eb1626c-kube-api-access-n8fx9\") pod \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085511 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data\") pod \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085551 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-combined-ca-bundle\") pod \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085578 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-vencrypt-tls-certs\") pod \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085646 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-combined-ca-bundle\") pod \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085679 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89r6q\" (UniqueName: \"kubernetes.io/projected/c986e5b7-9864-46ae-9c99-9072e9c6ce34-kube-api-access-89r6q\") pod \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085757 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-nova-novncproxy-tls-certs\") pod \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085782 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-config-data\") pod \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\" (UID: \"3feaa8e4-537c-44d2-ae92-ba999eb1626c\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085817 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68frq\" (UniqueName: \"kubernetes.io/projected/a9ee9f9f-66be-467a-a982-9ede1e1b682c-kube-api-access-68frq\") pod \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085863 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c986e5b7-9864-46ae-9c99-9072e9c6ce34-logs\") pod \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085899 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data-custom\") pod \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085941 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9ee9f9f-66be-467a-a982-9ede1e1b682c-logs\") pod \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\" (UID: \"a9ee9f9f-66be-467a-a982-9ede1e1b682c\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.085981 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data\") pod \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.086054 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data-custom\") pod \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.093096 4799 scope.go:117] "RemoveContainer" containerID="2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c" Nov 24 07:09:14 crc kubenswrapper[4799]: E1124 07:09:14.095007 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c\": container with ID starting with 2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c not found: ID does not exist" containerID="2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.095043 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c"} err="failed to get container status \"2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c\": rpc error: code = NotFound desc = could not find container \"2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c\": container with ID starting with 2a86e9bba871bb091ccac2bb11cf1604fa28179f0c69d9f4dc65086c3dba1d0c not found: ID does not exist" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.095072 4799 scope.go:117] "RemoveContainer" containerID="477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.097434 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3feaa8e4-537c-44d2-ae92-ba999eb1626c-kube-api-access-n8fx9" (OuterVolumeSpecName: "kube-api-access-n8fx9") pod "3feaa8e4-537c-44d2-ae92-ba999eb1626c" (UID: "3feaa8e4-537c-44d2-ae92-ba999eb1626c"). InnerVolumeSpecName "kube-api-access-n8fx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: E1124 07:09:14.097887 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742\": container with ID starting with 477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742 not found: ID does not exist" containerID="477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.097921 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742"} err="failed to get container status \"477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742\": rpc error: code = NotFound desc = could not find container \"477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742\": container with ID starting with 477432c6d5011c603c8a5cabd782d966a78d7de7205c2a16b5d35aaf713c9742 not found: ID does not exist" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.097949 4799 scope.go:117] "RemoveContainer" containerID="a9ca7d86df080d4b606da7e5a800c32f1929eb3f6ccde8a25eb20a206bdf9fda" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.103683 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-fxr57"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.107030 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.109024 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9ee9f9f-66be-467a-a982-9ede1e1b682c-logs" (OuterVolumeSpecName: "logs") pod "a9ee9f9f-66be-467a-a982-9ede1e1b682c" (UID: "a9ee9f9f-66be-467a-a982-9ede1e1b682c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.120390 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.138311 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a9ee9f9f-66be-467a-a982-9ede1e1b682c" (UID: "a9ee9f9f-66be-467a-a982-9ede1e1b682c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.142232 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c986e5b7-9864-46ae-9c99-9072e9c6ce34-logs" (OuterVolumeSpecName: "logs") pod "c986e5b7-9864-46ae-9c99-9072e9c6ce34" (UID: "c986e5b7-9864-46ae-9c99-9072e9c6ce34"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.143895 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c986e5b7-9864-46ae-9c99-9072e9c6ce34-kube-api-access-89r6q" (OuterVolumeSpecName: "kube-api-access-89r6q") pod "c986e5b7-9864-46ae-9c99-9072e9c6ce34" (UID: "c986e5b7-9864-46ae-9c99-9072e9c6ce34"). InnerVolumeSpecName "kube-api-access-89r6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.143945 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c986e5b7-9864-46ae-9c99-9072e9c6ce34" (UID: "c986e5b7-9864-46ae-9c99-9072e9c6ce34"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.169681 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ee9f9f-66be-467a-a982-9ede1e1b682c-kube-api-access-68frq" (OuterVolumeSpecName: "kube-api-access-68frq") pod "a9ee9f9f-66be-467a-a982-9ede1e1b682c" (UID: "a9ee9f9f-66be-467a-a982-9ede1e1b682c"). InnerVolumeSpecName "kube-api-access-68frq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.159879 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi722e-account-delete-th6f8"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.193892 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w69r4\" (UniqueName: \"kubernetes.io/projected/0d55de08-d6b7-44f6-8e96-44fcd8999000-kube-api-access-w69r4\") pod \"0d55de08-d6b7-44f6-8e96-44fcd8999000\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.194023 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-config-data-default\") pod \"844d53db-a457-4153-bdb4-d557ba34c099\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.194058 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-kolla-config\") pod \"844d53db-a457-4153-bdb4-d557ba34c099\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.194104 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-galera-tls-certs\") pod \"844d53db-a457-4153-bdb4-d557ba34c099\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.194147 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-operator-scripts\") pod \"844d53db-a457-4153-bdb4-d557ba34c099\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.194186 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-config-data\") pod \"0d55de08-d6b7-44f6-8e96-44fcd8999000\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.194217 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-combined-ca-bundle\") pod \"0d55de08-d6b7-44f6-8e96-44fcd8999000\" (UID: \"0d55de08-d6b7-44f6-8e96-44fcd8999000\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.194357 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-combined-ca-bundle\") pod \"844d53db-a457-4153-bdb4-d557ba34c099\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.194391 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd7rz\" (UniqueName: \"kubernetes.io/projected/844d53db-a457-4153-bdb4-d557ba34c099-kube-api-access-gd7rz\") pod \"844d53db-a457-4153-bdb4-d557ba34c099\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.194445 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/844d53db-a457-4153-bdb4-d557ba34c099-config-data-generated\") pod \"844d53db-a457-4153-bdb4-d557ba34c099\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.194501 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"844d53db-a457-4153-bdb4-d557ba34c099\" (UID: \"844d53db-a457-4153-bdb4-d557ba34c099\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.202314 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/844d53db-a457-4153-bdb4-d557ba34c099-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "844d53db-a457-4153-bdb4-d557ba34c099" (UID: "844d53db-a457-4153-bdb4-d557ba34c099"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.203620 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "844d53db-a457-4153-bdb4-d557ba34c099" (UID: "844d53db-a457-4153-bdb4-d557ba34c099"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.204275 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "844d53db-a457-4153-bdb4-d557ba34c099" (UID: "844d53db-a457-4153-bdb4-d557ba34c099"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.204774 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "844d53db-a457-4153-bdb4-d557ba34c099" (UID: "844d53db-a457-4153-bdb4-d557ba34c099"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.205841 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89r6q\" (UniqueName: \"kubernetes.io/projected/c986e5b7-9864-46ae-9c99-9072e9c6ce34-kube-api-access-89r6q\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.206104 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68frq\" (UniqueName: \"kubernetes.io/projected/a9ee9f9f-66be-467a-a982-9ede1e1b682c-kube-api-access-68frq\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.206120 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c986e5b7-9864-46ae-9c99-9072e9c6ce34-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.206145 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.206158 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9ee9f9f-66be-467a-a982-9ede1e1b682c-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.206171 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.206186 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8fx9\" (UniqueName: \"kubernetes.io/projected/3feaa8e4-537c-44d2-ae92-ba999eb1626c-kube-api-access-n8fx9\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.254165 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.275486 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d55de08-d6b7-44f6-8e96-44fcd8999000-kube-api-access-w69r4" (OuterVolumeSpecName: "kube-api-access-w69r4") pod "0d55de08-d6b7-44f6-8e96-44fcd8999000" (UID: "0d55de08-d6b7-44f6-8e96-44fcd8999000"). InnerVolumeSpecName "kube-api-access-w69r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.277145 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/844d53db-a457-4153-bdb4-d557ba34c099-kube-api-access-gd7rz" (OuterVolumeSpecName: "kube-api-access-gd7rz") pod "844d53db-a457-4153-bdb4-d557ba34c099" (UID: "844d53db-a457-4153-bdb4-d557ba34c099"). InnerVolumeSpecName "kube-api-access-gd7rz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.282179 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.303174 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9ee9f9f-66be-467a-a982-9ede1e1b682c" (UID: "a9ee9f9f-66be-467a-a982-9ede1e1b682c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.306999 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-config-data" (OuterVolumeSpecName: "config-data") pod "3feaa8e4-537c-44d2-ae92-ba999eb1626c" (UID: "3feaa8e4-537c-44d2-ae92-ba999eb1626c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.308589 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd7rz\" (UniqueName: \"kubernetes.io/projected/844d53db-a457-4153-bdb4-d557ba34c099-kube-api-access-gd7rz\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.308658 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.308676 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/844d53db-a457-4153-bdb4-d557ba34c099-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.308735 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w69r4\" (UniqueName: \"kubernetes.io/projected/0d55de08-d6b7-44f6-8e96-44fcd8999000-kube-api-access-w69r4\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.310160 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.310189 4799 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.310206 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/844d53db-a457-4153-bdb4-d557ba34c099-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.310222 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.333289 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "mysql-db") pod "844d53db-a457-4153-bdb4-d557ba34c099" (UID: "844d53db-a457-4153-bdb4-d557ba34c099"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.340527 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d55de08-d6b7-44f6-8e96-44fcd8999000" (UID: "0d55de08-d6b7-44f6-8e96-44fcd8999000"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.353645 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-config-data" (OuterVolumeSpecName: "config-data") pod "0d55de08-d6b7-44f6-8e96-44fcd8999000" (UID: "0d55de08-d6b7-44f6-8e96-44fcd8999000"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.363178 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data" (OuterVolumeSpecName: "config-data") pod "a9ee9f9f-66be-467a-a982-9ede1e1b682c" (UID: "a9ee9f9f-66be-467a-a982-9ede1e1b682c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.380441 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "3feaa8e4-537c-44d2-ae92-ba999eb1626c" (UID: "3feaa8e4-537c-44d2-ae92-ba999eb1626c"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.384609 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data" (OuterVolumeSpecName: "config-data") pod "c986e5b7-9864-46ae-9c99-9072e9c6ce34" (UID: "c986e5b7-9864-46ae-9c99-9072e9c6ce34"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.386068 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron2d2f-account-delete-5c2sd"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.399681 4799 scope.go:117] "RemoveContainer" containerID="7d643d949a4bb973305b6b986dbd4165fc4a1f0f0846279bc0b5ddf8929ad32b" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.404043 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican44f2-account-delete-2hm9q"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.409071 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement2704-account-delete-ck9t2"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.414096 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.414233 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d55de08-d6b7-44f6-8e96-44fcd8999000-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.414315 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ee9f9f-66be-467a-a982-9ede1e1b682c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.414411 4799 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.414594 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.418081 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.414942 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell000c8-account-delete-n5mhs"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.414868 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "3feaa8e4-537c-44d2-ae92-ba999eb1626c" (UID: "3feaa8e4-537c-44d2-ae92-ba999eb1626c"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.420865 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell125d6-account-delete-cw4x8"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.422572 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell000c8-account-delete-n5mhs" event={"ID":"87439e47-6c84-4467-a49b-03ed6e86a9eb","Type":"ContainerStarted","Data":"f56d12bad72bbd48e989299f0bd988a09be84a662c5b6cba9efd1892f6217316"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.428469 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-9445d5865-fhgvf"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.429675 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi722e-account-delete-th6f8" event={"ID":"63757cd4-2ed0-4423-b466-4dfac76e3ac4","Type":"ContainerStarted","Data":"0de054a9941a3f83dadfba70e8e2636c51690ca41971fa85674b31ef65e3d353"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.433164 4799 generic.go:334] "Generic (PLEG): container finished" podID="d9064d19-6806-4e78-beb4-cda924fe3010" containerID="d59b7428d4e280f7869415161a1fab26ded6505c40670360f51e2d5cf153d262" exitCode=0 Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.433313 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderbbb1-account-delete-tjfp2" event={"ID":"d9064d19-6806-4e78-beb4-cda924fe3010","Type":"ContainerDied","Data":"d59b7428d4e280f7869415161a1fab26ded6505c40670360f51e2d5cf153d262"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.435092 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-9445d5865-fhgvf"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.438678 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.438889 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fbb66bcc8-62525" event={"ID":"c986e5b7-9864-46ae-9c99-9072e9c6ce34","Type":"ContainerDied","Data":"b59051fadd3bd06a2fad5470bc4264fd6214b1856014a42cba1775ec385c56ad"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.440391 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement2704-account-delete-ck9t2" event={"ID":"b6f22313-6d9a-4b35-b892-695903d354b1","Type":"ContainerStarted","Data":"9a2d2c63d77347f5059d85279c8ba9c130ff9e52541316e83e9fd29653b25b04"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.442266 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron2d2f-account-delete-5c2sd" event={"ID":"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee","Type":"ContainerStarted","Data":"3e6f01b2ffc127b14096eb3eb3ce0f7eb3f9e8f18f4b5d85f347feda67faa3e4"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.451493 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7dd777b755-wgf6d" event={"ID":"a9ee9f9f-66be-467a-a982-9ede1e1b682c","Type":"ContainerDied","Data":"953ffb6c5a3fad664636d7f6e9310636d909840ca49d65fdc75eb8af5c39dd92"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.451590 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7dd777b755-wgf6d" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.457350 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3feaa8e4-537c-44d2-ae92-ba999eb1626c" (UID: "3feaa8e4-537c-44d2-ae92-ba999eb1626c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.457748 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "844d53db-a457-4153-bdb4-d557ba34c099" (UID: "844d53db-a457-4153-bdb4-d557ba34c099"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.462571 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3feaa8e4-537c-44d2-ae92-ba999eb1626c","Type":"ContainerDied","Data":"57719c24ba2aa35603649c3f2a0655efe9f343e4b6858acf1844fc6dd510aac8"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.462596 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.470511 4799 generic.go:334] "Generic (PLEG): container finished" podID="ad75f933-ab36-4c9d-857d-c2cc11702dd3" containerID="0b036e7485bdbc7a10091975ed6eae68dfe9ad069cb89ee4aca3cd050c565d50" exitCode=0 Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.470612 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance00d0-account-delete-kbtpl" event={"ID":"ad75f933-ab36-4c9d-857d-c2cc11702dd3","Type":"ContainerDied","Data":"0b036e7485bdbc7a10091975ed6eae68dfe9ad069cb89ee4aca3cd050c565d50"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.472326 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican44f2-account-delete-2hm9q" event={"ID":"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a","Type":"ContainerStarted","Data":"5b4890d36b23a1d16cc862079b4d9b8b999f8afb812820c384015d6702072fbf"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.497480 4799 generic.go:334] "Generic (PLEG): container finished" podID="0d55de08-d6b7-44f6-8e96-44fcd8999000" containerID="0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac" exitCode=0 Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.497544 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d55de08-d6b7-44f6-8e96-44fcd8999000","Type":"ContainerDied","Data":"0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.497568 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d55de08-d6b7-44f6-8e96-44fcd8999000","Type":"ContainerDied","Data":"97c4e8b888ca0aacccacf821815670ffc6cf28575c35f660f18b84923156e041"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.497633 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.499527 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.499838 4799 scope.go:117] "RemoveContainer" containerID="1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.519972 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c986e5b7-9864-46ae-9c99-9072e9c6ce34" (UID: "c986e5b7-9864-46ae-9c99-9072e9c6ce34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.520139 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-combined-ca-bundle\") pod \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\" (UID: \"c986e5b7-9864-46ae-9c99-9072e9c6ce34\") " Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.521058 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.521080 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.521089 4799 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3feaa8e4-537c-44d2-ae92-ba999eb1626c-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.521099 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: W1124 07:09:14.521173 4799 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/c986e5b7-9864-46ae-9c99-9072e9c6ce34/volumes/kubernetes.io~secret/combined-ca-bundle Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.521186 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c986e5b7-9864-46ae-9c99-9072e9c6ce34" (UID: "c986e5b7-9864-46ae-9c99-9072e9c6ce34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.526512 4799 generic.go:334] "Generic (PLEG): container finished" podID="844d53db-a457-4153-bdb4-d557ba34c099" containerID="57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e" exitCode=0 Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.526823 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.527742 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"844d53db-a457-4153-bdb4-d557ba34c099","Type":"ContainerDied","Data":"57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.527784 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"844d53db-a457-4153-bdb4-d557ba34c099","Type":"ContainerDied","Data":"46cd63af2e8051dfb8835b25509abf388109dc011bc4a05c232bafe7a0f0c1e5"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.533451 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell125d6-account-delete-cw4x8" event={"ID":"d4a27166-5918-4ee7-8569-a9f58914d7be","Type":"ContainerStarted","Data":"51e4e0483f26c31669ce075027014e0a4faa36f8f873839a0c1c1a89d908f12d"} Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.536881 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7dd777b755-wgf6d"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.566041 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-7dd777b755-wgf6d"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.579794 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "844d53db-a457-4153-bdb4-d557ba34c099" (UID: "844d53db-a457-4153-bdb4-d557ba34c099"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.579893 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.591357 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="f643414c-72f0-4e7c-b804-20a56f3207ab" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.166:8776/healthcheck\": read tcp 10.217.0.2:45546->10.217.0.166:8776: read: connection reset by peer" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.591662 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.599215 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.604741 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.624575 4799 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/844d53db-a457-4153-bdb4-d557ba34c099-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.624615 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c986e5b7-9864-46ae-9c99-9072e9c6ce34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.634908 4799 scope.go:117] "RemoveContainer" containerID="1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259" Nov 24 07:09:14 crc kubenswrapper[4799]: E1124 07:09:14.636363 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259\": container with ID starting with 1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259 not found: ID does not exist" containerID="1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.636402 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259"} err="failed to get container status \"1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259\": rpc error: code = NotFound desc = could not find container \"1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259\": container with ID starting with 1c485376ad18a47d3c557bb0368cf4ad88684a1a55d72de0215e6c1d963b3259 not found: ID does not exist" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.636429 4799 scope.go:117] "RemoveContainer" containerID="38b78508876ab5b361ddd58cb9e13822b959fef1c0bad5d5fe801de1bd871409" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.772414 4799 scope.go:117] "RemoveContainer" containerID="a566d8f356e5254e7334281ca19afeaafb46a936eff3f95e9bb6c95fec8e862f" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.787287 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6fbb66bcc8-62525"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.796573 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-6fbb66bcc8-62525"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.830235 4799 scope.go:117] "RemoveContainer" containerID="f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.881659 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.882398 4799 scope.go:117] "RemoveContainer" containerID="2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.887345 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.909129 4799 scope.go:117] "RemoveContainer" containerID="f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b" Nov 24 07:09:14 crc kubenswrapper[4799]: E1124 07:09:14.909585 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b\": container with ID starting with f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b not found: ID does not exist" containerID="f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.909629 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b"} err="failed to get container status \"f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b\": rpc error: code = NotFound desc = could not find container \"f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b\": container with ID starting with f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b not found: ID does not exist" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.909661 4799 scope.go:117] "RemoveContainer" containerID="2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a" Nov 24 07:09:14 crc kubenswrapper[4799]: E1124 07:09:14.912767 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a\": container with ID starting with 2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a not found: ID does not exist" containerID="2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.912793 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a"} err="failed to get container status \"2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a\": rpc error: code = NotFound desc = could not find container \"2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a\": container with ID starting with 2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a not found: ID does not exist" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.912812 4799 scope.go:117] "RemoveContainer" containerID="f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.913223 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b"} err="failed to get container status \"f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b\": rpc error: code = NotFound desc = could not find container \"f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b\": container with ID starting with f10fe12f106216420cf38c1e98a4fd4c0bd2e416f5c8bcbcbf2aae357e0c634b not found: ID does not exist" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.913242 4799 scope.go:117] "RemoveContainer" containerID="2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.916067 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a"} err="failed to get container status \"2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a\": rpc error: code = NotFound desc = could not find container \"2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a\": container with ID starting with 2d2fd14341731aa41f40e32f1cbdb8cbaac39f16c2768ea9beaf697ae8970a1a not found: ID does not exist" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.916102 4799 scope.go:117] "RemoveContainer" containerID="90b897d4e24c66a97b38d28448dae228ba7620bc06b2cd553b7c3edcb5ea7710" Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.986902 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.987557 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="ceilometer-central-agent" containerID="cri-o://4b1017dff22faa9e7cf62359f3de7ceea284aa7bf591e9d872e15ff649d22d96" gracePeriod=30 Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.987698 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="proxy-httpd" containerID="cri-o://dc5fdc4d831816407f510ddb2a4fc3bcb5bbc530983dfb627e98b50af968d2bd" gracePeriod=30 Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.987730 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="sg-core" containerID="cri-o://6b464ae1867d289634184c0a2a89b04d292707252e13139c622836b8234cc5e5" gracePeriod=30 Nov 24 07:09:14 crc kubenswrapper[4799]: I1124 07:09:14.987899 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="ceilometer-notification-agent" containerID="cri-o://247ac76988e78afd0fb67b14cc28cf0d059ad2c2f3abbf816f4a4d4fb2204cb3" gracePeriod=30 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.018721 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.019009 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="3cd44457-1418-4788-a029-82c54c917997" containerName="kube-state-metrics" containerID="cri-o://e7c302a760dcca397dd206e8bf79d8516ca32ba49a1ef37a43e67e1893d9366d" gracePeriod=30 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.154250 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.154479 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="b0361baf-d0c3-430f-87e3-c011f869a2b3" containerName="memcached" containerID="cri-o://4bcf19e3b710fbf877a92e707435d95962e55b4801e15a2e5b43f689d28b8bba" gracePeriod=30 Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.245181 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.245543 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-bclrf"] Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.248973 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.264152 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-bclrf"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.268927 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-h5vcf"] Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.277982 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.278054 4799 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server" Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.278282 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.278542 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-h5vcf"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.293530 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7b94787bcf-mts4z"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.293802 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-7b94787bcf-mts4z" podUID="7603a96c-19e2-48ce-9c5f-69eeacc8cd85" containerName="keystone-api" containerID="cri-o://29a4295f9247968819fb9ab51fb68ba67eeb2fb13b12fb980c93984bfc24c9e9" gracePeriod=30 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.298978 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.348210 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.355094 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-2dz2n" podUID="dd32bea6-7ad9-4182-bc59-5644efe24ce1" containerName="ovn-controller" probeResult="failure" output=< Nov 24 07:09:15 crc kubenswrapper[4799]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Nov 24 07:09:15 crc kubenswrapper[4799]: > Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.355176 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.355205 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovs-vswitchd" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.365316 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-wsc8j"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.392941 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-wsc8j"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.431814 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-2704-account-create-mt2ps"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.445663 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-2704-account-create-mt2ps"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.455121 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": read tcp 10.217.0.2:59948->10.217.0.204:8775: read: connection reset by peer" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.455808 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": read tcp 10.217.0.2:59944->10.217.0.204:8775: read: connection reset by peer" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.460668 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.199:3000/\": read tcp 10.217.0.2:57504->10.217.0.199:3000: read: connection reset by peer" Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.489188 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="192c554265f38b0137575a7bb2ad8ef843ac317c21ba65066896ec0465f061f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.496509 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="192c554265f38b0137575a7bb2ad8ef843ac317c21ba65066896ec0465f061f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.498979 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="192c554265f38b0137575a7bb2ad8ef843ac317c21ba65066896ec0465f061f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.499018 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="168550d3-6597-41ff-adf4-d20f190c7a02" containerName="nova-cell1-conductor-conductor" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.500336 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement2704-account-delete-ck9t2"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.510783 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-ljnx2"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.520054 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-ljnx2"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.525781 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-cabe-account-create-zjzwn"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.536816 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-cabe-account-create-zjzwn"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.586378 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-z877t"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.593487 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-z877t"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.605684 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-00c8-account-create-w4gf7"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.623220 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell000c8-account-delete-n5mhs"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.670663 4799 generic.go:334] "Generic (PLEG): container finished" podID="f643414c-72f0-4e7c-b804-20a56f3207ab" containerID="30959909b72adeb936fc238d713e097048621efd99d9e5661c63d6da5128b9e8" exitCode=0 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.696920 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09082d4e-0623-40c5-b584-1eb58985cfc0" path="/var/lib/kubelet/pods/09082d4e-0623-40c5-b584-1eb58985cfc0/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.699426 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d55de08-d6b7-44f6-8e96-44fcd8999000" path="/var/lib/kubelet/pods/0d55de08-d6b7-44f6-8e96-44fcd8999000/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.700541 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" path="/var/lib/kubelet/pods/2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.701216 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="997ffb9e-cc6a-4fca-91d9-eb17d711b135" containerName="galera" containerID="cri-o://63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174" gracePeriod=30 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.707422 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32a13216-4d25-462a-b55b-59602988c3c4" path="/var/lib/kubelet/pods/32a13216-4d25-462a-b55b-59602988c3c4/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.708359 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3feaa8e4-537c-44d2-ae92-ba999eb1626c" path="/var/lib/kubelet/pods/3feaa8e4-537c-44d2-ae92-ba999eb1626c/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.712267 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed" path="/var/lib/kubelet/pods/558a22c8-97cd-48ab-b7f5-fdd5ea96a5ed/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.712940 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="844d53db-a457-4153-bdb4-d557ba34c099" path="/var/lib/kubelet/pods/844d53db-a457-4153-bdb4-d557ba34c099/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.714827 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98d23df1-e416-488c-b1bc-d075d5b63fb3" path="/var/lib/kubelet/pods/98d23df1-e416-488c-b1bc-d075d5b63fb3/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.727787 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ee9f9f-66be-467a-a982-9ede1e1b682c" path="/var/lib/kubelet/pods/a9ee9f9f-66be-467a-a982-9ede1e1b682c/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.728369 4799 generic.go:334] "Generic (PLEG): container finished" podID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerID="6b464ae1867d289634184c0a2a89b04d292707252e13139c622836b8234cc5e5" exitCode=2 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.728979 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b639ad0e-8438-4a36-a776-ee53b6f0c080" path="/var/lib/kubelet/pods/b639ad0e-8438-4a36-a776-ee53b6f0c080/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.753218 4799 generic.go:334] "Generic (PLEG): container finished" podID="d3f86cb1-3f0e-4da8-82dd-e00522d4ceee" containerID="7d78517d5fccdbbf385d41ea0be66f81ad0ecb4529e90eaf0d139df4336fa2b1" exitCode=0 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.769387 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbfb60f9-6e71-4b22-9d74-cab607a00c20" path="/var/lib/kubelet/pods/bbfb60f9-6e71-4b22-9d74-cab607a00c20/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.770082 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3d5dc9f-307c-4217-a7df-c1b6551441a5" path="/var/lib/kubelet/pods/c3d5dc9f-307c-4217-a7df-c1b6551441a5/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.770632 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c986e5b7-9864-46ae-9c99-9072e9c6ce34" path="/var/lib/kubelet/pods/c986e5b7-9864-46ae-9c99-9072e9c6ce34/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.774426 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce1d3095-b344-4e6c-b49a-d1de3a7260ee" path="/var/lib/kubelet/pods/ce1d3095-b344-4e6c-b49a-d1de3a7260ee/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.775822 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7128a3b-75e5-49c4-b25c-8d2a643d65cb" path="/var/lib/kubelet/pods/d7128a3b-75e5-49c4-b25c-8d2a643d65cb/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.776382 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100" path="/var/lib/kubelet/pods/e4ebd6a0-1d89-4d6f-948f-2f1a1a7a9100/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.783861 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe55fcc2-0636-4524-a2f5-8127ab456570" path="/var/lib/kubelet/pods/fe55fcc2-0636-4524-a2f5-8127ab456570/volumes" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.784523 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-00c8-account-create-w4gf7"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.784558 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell125d6-account-delete-cw4x8" event={"ID":"d4a27166-5918-4ee7-8569-a9f58914d7be","Type":"ContainerStarted","Data":"36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.784584 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f643414c-72f0-4e7c-b804-20a56f3207ab","Type":"ContainerDied","Data":"30959909b72adeb936fc238d713e097048621efd99d9e5661c63d6da5128b9e8"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.784603 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f643414c-72f0-4e7c-b804-20a56f3207ab","Type":"ContainerDied","Data":"6132ba1992328a5bbcc5e80d2617c0eda719326e31d71528f2970cfeb453f488"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.784618 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6132ba1992328a5bbcc5e80d2617c0eda719326e31d71528f2970cfeb453f488" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.784629 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-mwx4g"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.784643 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f97d9a56-cb78-4da9-b832-0028b6e45895","Type":"ContainerDied","Data":"6b464ae1867d289634184c0a2a89b04d292707252e13139c622836b8234cc5e5"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.784657 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-mwx4g"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.784674 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron2d2f-account-delete-5c2sd" event={"ID":"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee","Type":"ContainerDied","Data":"7d78517d5fccdbbf385d41ea0be66f81ad0ecb4529e90eaf0d139df4336fa2b1"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.788352 4799 generic.go:334] "Generic (PLEG): container finished" podID="897d0d90-bf27-4d5b-b136-863379656f34" containerID="62f862c6ff3823a347d374cd9375930cec2ce5b254235146abb2a0f3a2a1a61c" exitCode=0 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.788435 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65f6685878-7bjdh" event={"ID":"897d0d90-bf27-4d5b-b136-863379656f34","Type":"ContainerDied","Data":"62f862c6ff3823a347d374cd9375930cec2ce5b254235146abb2a0f3a2a1a61c"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.802545 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi722e-account-delete-th6f8"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.807547 4799 generic.go:334] "Generic (PLEG): container finished" podID="1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" containerID="183c3c900b89e4d354f72e7ee6593a3ca4a44de57d248b7f61ac4959fe51c8ef" exitCode=0 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.807628 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf","Type":"ContainerDied","Data":"183c3c900b89e4d354f72e7ee6593a3ca4a44de57d248b7f61ac4959fe51c8ef"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.810064 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell000c8-account-delete-n5mhs" event={"ID":"87439e47-6c84-4467-a49b-03ed6e86a9eb","Type":"ContainerStarted","Data":"44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.810267 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-722e-account-create-5cz9z"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.811321 4799 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell000c8-account-delete-n5mhs" secret="" err="secret \"galera-openstack-dockercfg-hptzz\" not found" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.817315 4799 generic.go:334] "Generic (PLEG): container finished" podID="3cd44457-1418-4788-a029-82c54c917997" containerID="e7c302a760dcca397dd206e8bf79d8516ca32ba49a1ef37a43e67e1893d9366d" exitCode=2 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.817383 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3cd44457-1418-4788-a029-82c54c917997","Type":"ContainerDied","Data":"e7c302a760dcca397dd206e8bf79d8516ca32ba49a1ef37a43e67e1893d9366d"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.823040 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican44f2-account-delete-2hm9q" event={"ID":"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a","Type":"ContainerStarted","Data":"16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.823779 4799 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican44f2-account-delete-2hm9q" secret="" err="secret \"galera-openstack-dockercfg-hptzz\" not found" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.826001 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-722e-account-create-5cz9z"] Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.838224 4799 generic.go:334] "Generic (PLEG): container finished" podID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerID="f6b04750444de5a3182982a1b53fc81a035e9a9f2eb6df3d21f831648f26fc49" exitCode=0 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.838299 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fe03a76-2eb9-4728-9396-9d2dc1e743af","Type":"ContainerDied","Data":"f6b04750444de5a3182982a1b53fc81a035e9a9f2eb6df3d21f831648f26fc49"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.874822 4799 generic.go:334] "Generic (PLEG): container finished" podID="e9a036d8-ce56-4654-a782-15cd18730378" containerID="e97f64b93e8c87fb3e9989776d9be9cfbdd54d15ade6362a154f53220314a180" exitCode=0 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.875002 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e9a036d8-ce56-4654-a782-15cd18730378","Type":"ContainerDied","Data":"e97f64b93e8c87fb3e9989776d9be9cfbdd54d15ade6362a154f53220314a180"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.901979 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement2704-account-delete-ck9t2" event={"ID":"b6f22313-6d9a-4b35-b892-695903d354b1","Type":"ContainerStarted","Data":"fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.902478 4799 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/placement2704-account-delete-ck9t2" secret="" err="secret \"galera-openstack-dockercfg-hptzz\" not found" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.909634 4799 generic.go:334] "Generic (PLEG): container finished" podID="63757cd4-2ed0-4423-b466-4dfac76e3ac4" containerID="ecec47ab274347df90c4772ded8aa7a5894d4529edf5f63f16082a56ca0c1da3" exitCode=0 Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.909823 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi722e-account-delete-th6f8" event={"ID":"63757cd4-2ed0-4423-b466-4dfac76e3ac4","Type":"ContainerDied","Data":"ecec47ab274347df90c4772ded8aa7a5894d4529edf5f63f16082a56ca0c1da3"} Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.939142 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell000c8-account-delete-n5mhs" podStartSLOduration=5.939121324 podStartE2EDuration="5.939121324s" podCreationTimestamp="2025-11-24 07:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:09:15.82832476 +0000 UTC m=+1301.484307234" watchObservedRunningTime="2025-11-24 07:09:15.939121324 +0000 UTC m=+1301.595103798" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.949306 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican44f2-account-delete-2hm9q" podStartSLOduration=5.949289652 podStartE2EDuration="5.949289652s" podCreationTimestamp="2025-11-24 07:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:09:15.841665429 +0000 UTC m=+1301.497647903" watchObservedRunningTime="2025-11-24 07:09:15.949289652 +0000 UTC m=+1301.605272126" Nov 24 07:09:15 crc kubenswrapper[4799]: I1124 07:09:15.953389 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement2704-account-delete-ck9t2" podStartSLOduration=5.953362918 podStartE2EDuration="5.953362918s" podCreationTimestamp="2025-11-24 07:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:09:15.928224414 +0000 UTC m=+1301.584206898" watchObservedRunningTime="2025-11-24 07:09:15.953362918 +0000 UTC m=+1301.609345392" Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.966920 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.967005 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts podName:87439e47-6c84-4467-a49b-03ed6e86a9eb nodeName:}" failed. No retries permitted until 2025-11-24 07:09:16.466988625 +0000 UTC m=+1302.122971099 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts") pod "novacell000c8-account-delete-n5mhs" (UID: "87439e47-6c84-4467-a49b-03ed6e86a9eb") : configmap "openstack-scripts" not found Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.967209 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:15 crc kubenswrapper[4799]: E1124 07:09:15.967286 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts podName:d13b66c3-1e31-418e-a0ff-0d1bf7a5980a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:16.467266132 +0000 UTC m=+1302.123248606 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts") pod "barbican44f2-account-delete-2hm9q" (UID: "d13b66c3-1e31-418e-a0ff-0d1bf7a5980a") : configmap "openstack-scripts" not found Nov 24 07:09:16 crc kubenswrapper[4799]: E1124 07:09:16.068927 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:16 crc kubenswrapper[4799]: E1124 07:09:16.068985 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts podName:b6f22313-6d9a-4b35-b892-695903d354b1 nodeName:}" failed. No retries permitted until 2025-11-24 07:09:16.568972028 +0000 UTC m=+1302.224954492 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts") pod "placement2704-account-delete-ck9t2" (UID: "b6f22313-6d9a-4b35-b892-695903d354b1") : configmap "openstack-scripts" not found Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.131930 4799 scope.go:117] "RemoveContainer" containerID="d87462cbe65628447244d9ea1c17fb1adb8cfc23702f9f99482cba9b90403d7c" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.147494 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.189486 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.193528 4799 scope.go:117] "RemoveContainer" containerID="f9a800a0af5f1a9b50d8af38a76b70c4541840dcf2598a2012116de5e32ec921" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.199925 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.268879 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.276699 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data\") pod \"f643414c-72f0-4e7c-b804-20a56f3207ab\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.276756 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-scripts\") pod \"f643414c-72f0-4e7c-b804-20a56f3207ab\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.276808 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dk9x\" (UniqueName: \"kubernetes.io/projected/f643414c-72f0-4e7c-b804-20a56f3207ab-kube-api-access-9dk9x\") pod \"f643414c-72f0-4e7c-b804-20a56f3207ab\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.276837 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data-custom\") pod \"f643414c-72f0-4e7c-b804-20a56f3207ab\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.276978 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f643414c-72f0-4e7c-b804-20a56f3207ab-logs\") pod \"f643414c-72f0-4e7c-b804-20a56f3207ab\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.277085 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-combined-ca-bundle\") pod \"f643414c-72f0-4e7c-b804-20a56f3207ab\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.277114 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-internal-tls-certs\") pod \"f643414c-72f0-4e7c-b804-20a56f3207ab\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.277150 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-public-tls-certs\") pod \"f643414c-72f0-4e7c-b804-20a56f3207ab\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.277334 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f643414c-72f0-4e7c-b804-20a56f3207ab-etc-machine-id\") pod \"f643414c-72f0-4e7c-b804-20a56f3207ab\" (UID: \"f643414c-72f0-4e7c-b804-20a56f3207ab\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.277961 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f643414c-72f0-4e7c-b804-20a56f3207ab-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f643414c-72f0-4e7c-b804-20a56f3207ab" (UID: "f643414c-72f0-4e7c-b804-20a56f3207ab"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.278705 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f643414c-72f0-4e7c-b804-20a56f3207ab-logs" (OuterVolumeSpecName: "logs") pod "f643414c-72f0-4e7c-b804-20a56f3207ab" (UID: "f643414c-72f0-4e7c-b804-20a56f3207ab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.293579 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f643414c-72f0-4e7c-b804-20a56f3207ab-kube-api-access-9dk9x" (OuterVolumeSpecName: "kube-api-access-9dk9x") pod "f643414c-72f0-4e7c-b804-20a56f3207ab" (UID: "f643414c-72f0-4e7c-b804-20a56f3207ab"). InnerVolumeSpecName "kube-api-access-9dk9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.294428 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-scripts" (OuterVolumeSpecName: "scripts") pod "f643414c-72f0-4e7c-b804-20a56f3207ab" (UID: "f643414c-72f0-4e7c-b804-20a56f3207ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.295017 4799 scope.go:117] "RemoveContainer" containerID="51856ae48de556e52c655ad669ed192facc4a302f9d9889f4ff2d00441ecb8fa" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.295360 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f643414c-72f0-4e7c-b804-20a56f3207ab" (UID: "f643414c-72f0-4e7c-b804-20a56f3207ab"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.336255 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f643414c-72f0-4e7c-b804-20a56f3207ab" (UID: "f643414c-72f0-4e7c-b804-20a56f3207ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.374347 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f643414c-72f0-4e7c-b804-20a56f3207ab" (UID: "f643414c-72f0-4e7c-b804-20a56f3207ab"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.385935 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-scripts\") pod \"e9a036d8-ce56-4654-a782-15cd18730378\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.385981 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68hmq\" (UniqueName: \"kubernetes.io/projected/3cd44457-1418-4788-a029-82c54c917997-kube-api-access-68hmq\") pod \"3cd44457-1418-4788-a029-82c54c917997\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386047 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-certs\") pod \"3cd44457-1418-4788-a029-82c54c917997\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386073 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzcck\" (UniqueName: \"kubernetes.io/projected/e9a036d8-ce56-4654-a782-15cd18730378-kube-api-access-kzcck\") pod \"e9a036d8-ce56-4654-a782-15cd18730378\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386098 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-public-tls-certs\") pod \"897d0d90-bf27-4d5b-b136-863379656f34\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386124 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"e9a036d8-ce56-4654-a782-15cd18730378\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386229 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-config-data\") pod \"e9a036d8-ce56-4654-a782-15cd18730378\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386281 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-logs\") pod \"e9a036d8-ce56-4654-a782-15cd18730378\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386307 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-public-tls-certs\") pod \"e9a036d8-ce56-4654-a782-15cd18730378\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386345 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-internal-tls-certs\") pod \"897d0d90-bf27-4d5b-b136-863379656f34\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386373 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-httpd-run\") pod \"e9a036d8-ce56-4654-a782-15cd18730378\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386392 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-combined-ca-bundle\") pod \"e9a036d8-ce56-4654-a782-15cd18730378\" (UID: \"e9a036d8-ce56-4654-a782-15cd18730378\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386432 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s88fq\" (UniqueName: \"kubernetes.io/projected/897d0d90-bf27-4d5b-b136-863379656f34-kube-api-access-s88fq\") pod \"897d0d90-bf27-4d5b-b136-863379656f34\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386459 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-config-data\") pod \"897d0d90-bf27-4d5b-b136-863379656f34\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386477 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/897d0d90-bf27-4d5b-b136-863379656f34-logs\") pod \"897d0d90-bf27-4d5b-b136-863379656f34\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386507 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-combined-ca-bundle\") pod \"897d0d90-bf27-4d5b-b136-863379656f34\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386531 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-combined-ca-bundle\") pod \"3cd44457-1418-4788-a029-82c54c917997\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386568 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-scripts\") pod \"897d0d90-bf27-4d5b-b136-863379656f34\" (UID: \"897d0d90-bf27-4d5b-b136-863379656f34\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.386595 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-config\") pod \"3cd44457-1418-4788-a029-82c54c917997\" (UID: \"3cd44457-1418-4788-a029-82c54c917997\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.387081 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.387095 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.387107 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f643414c-72f0-4e7c-b804-20a56f3207ab-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.387116 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.387126 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.387135 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dk9x\" (UniqueName: \"kubernetes.io/projected/f643414c-72f0-4e7c-b804-20a56f3207ab-kube-api-access-9dk9x\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.387145 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f643414c-72f0-4e7c-b804-20a56f3207ab-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.388093 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e9a036d8-ce56-4654-a782-15cd18730378" (UID: "e9a036d8-ce56-4654-a782-15cd18730378"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.389292 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/897d0d90-bf27-4d5b-b136-863379656f34-logs" (OuterVolumeSpecName: "logs") pod "897d0d90-bf27-4d5b-b136-863379656f34" (UID: "897d0d90-bf27-4d5b-b136-863379656f34"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.389323 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-logs" (OuterVolumeSpecName: "logs") pod "e9a036d8-ce56-4654-a782-15cd18730378" (UID: "e9a036d8-ce56-4654-a782-15cd18730378"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.390906 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9a036d8-ce56-4654-a782-15cd18730378-kube-api-access-kzcck" (OuterVolumeSpecName: "kube-api-access-kzcck") pod "e9a036d8-ce56-4654-a782-15cd18730378" (UID: "e9a036d8-ce56-4654-a782-15cd18730378"). InnerVolumeSpecName "kube-api-access-kzcck". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.390953 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data" (OuterVolumeSpecName: "config-data") pod "f643414c-72f0-4e7c-b804-20a56f3207ab" (UID: "f643414c-72f0-4e7c-b804-20a56f3207ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.392596 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-scripts" (OuterVolumeSpecName: "scripts") pod "e9a036d8-ce56-4654-a782-15cd18730378" (UID: "e9a036d8-ce56-4654-a782-15cd18730378"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.396442 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f643414c-72f0-4e7c-b804-20a56f3207ab" (UID: "f643414c-72f0-4e7c-b804-20a56f3207ab"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.403138 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "e9a036d8-ce56-4654-a782-15cd18730378" (UID: "e9a036d8-ce56-4654-a782-15cd18730378"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.403238 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-scripts" (OuterVolumeSpecName: "scripts") pod "897d0d90-bf27-4d5b-b136-863379656f34" (UID: "897d0d90-bf27-4d5b-b136-863379656f34"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.404918 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/897d0d90-bf27-4d5b-b136-863379656f34-kube-api-access-s88fq" (OuterVolumeSpecName: "kube-api-access-s88fq") pod "897d0d90-bf27-4d5b-b136-863379656f34" (UID: "897d0d90-bf27-4d5b-b136-863379656f34"). InnerVolumeSpecName "kube-api-access-s88fq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.432303 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd44457-1418-4788-a029-82c54c917997-kube-api-access-68hmq" (OuterVolumeSpecName: "kube-api-access-68hmq") pod "3cd44457-1418-4788-a029-82c54c917997" (UID: "3cd44457-1418-4788-a029-82c54c917997"). InnerVolumeSpecName "kube-api-access-68hmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.432430 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9a036d8-ce56-4654-a782-15cd18730378" (UID: "e9a036d8-ce56-4654-a782-15cd18730378"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.449990 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cd44457-1418-4788-a029-82c54c917997" (UID: "3cd44457-1418-4788-a029-82c54c917997"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.477961 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e9a036d8-ce56-4654-a782-15cd18730378" (UID: "e9a036d8-ce56-4654-a782-15cd18730378"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.477982 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "3cd44457-1418-4788-a029-82c54c917997" (UID: "3cd44457-1418-4788-a029-82c54c917997"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.478453 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "3cd44457-1418-4788-a029-82c54c917997" (UID: "3cd44457-1418-4788-a029-82c54c917997"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.489590 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s88fq\" (UniqueName: \"kubernetes.io/projected/897d0d90-bf27-4d5b-b136-863379656f34-kube-api-access-s88fq\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.489631 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/897d0d90-bf27-4d5b-b136-863379656f34-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490156 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490181 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490195 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490207 4799 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490221 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490233 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68hmq\" (UniqueName: \"kubernetes.io/projected/3cd44457-1418-4788-a029-82c54c917997-kube-api-access-68hmq\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490245 4799 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd44457-1418-4788-a029-82c54c917997-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490257 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzcck\" (UniqueName: \"kubernetes.io/projected/e9a036d8-ce56-4654-a782-15cd18730378-kube-api-access-kzcck\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490294 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490307 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f643414c-72f0-4e7c-b804-20a56f3207ab-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490321 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490332 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490344 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e9a036d8-ce56-4654-a782-15cd18730378-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.490355 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: E1124 07:09:16.489962 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:16 crc kubenswrapper[4799]: E1124 07:09:16.491380 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts podName:87439e47-6c84-4467-a49b-03ed6e86a9eb nodeName:}" failed. No retries permitted until 2025-11-24 07:09:17.491354165 +0000 UTC m=+1303.147336639 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts") pod "novacell000c8-account-delete-n5mhs" (UID: "87439e47-6c84-4467-a49b-03ed6e86a9eb") : configmap "openstack-scripts" not found Nov 24 07:09:16 crc kubenswrapper[4799]: E1124 07:09:16.490126 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:16 crc kubenswrapper[4799]: E1124 07:09:16.491610 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts podName:d13b66c3-1e31-418e-a0ff-0d1bf7a5980a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:17.491589291 +0000 UTC m=+1303.147571845 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts") pod "barbican44f2-account-delete-2hm9q" (UID: "d13b66c3-1e31-418e-a0ff-0d1bf7a5980a") : configmap "openstack-scripts" not found Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.497548 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-config-data" (OuterVolumeSpecName: "config-data") pod "897d0d90-bf27-4d5b-b136-863379656f34" (UID: "897d0d90-bf27-4d5b-b136-863379656f34"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.522975 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.563443 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-config-data" (OuterVolumeSpecName: "config-data") pod "e9a036d8-ce56-4654-a782-15cd18730378" (UID: "e9a036d8-ce56-4654-a782-15cd18730378"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.568085 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.595297 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.595330 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a036d8-ce56-4654-a782-15cd18730378-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.595341 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.595529 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "897d0d90-bf27-4d5b-b136-863379656f34" (UID: "897d0d90-bf27-4d5b-b136-863379656f34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: E1124 07:09:16.596157 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:16 crc kubenswrapper[4799]: E1124 07:09:16.596204 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts podName:b6f22313-6d9a-4b35-b892-695903d354b1 nodeName:}" failed. No retries permitted until 2025-11-24 07:09:17.59619012 +0000 UTC m=+1303.252172594 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts") pod "placement2704-account-delete-ck9t2" (UID: "b6f22313-6d9a-4b35-b892-695903d354b1") : configmap "openstack-scripts" not found Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.601530 4799 scope.go:117] "RemoveContainer" containerID="330d7990c1e4ba99f0cea6c8c2d506420220e3481e87800ebda2781888dd174a" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.616699 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "897d0d90-bf27-4d5b-b136-863379656f34" (UID: "897d0d90-bf27-4d5b-b136-863379656f34"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.644299 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "897d0d90-bf27-4d5b-b136-863379656f34" (UID: "897d0d90-bf27-4d5b-b136-863379656f34"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.700361 4799 scope.go:117] "RemoveContainer" containerID="0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.700951 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkwpb\" (UniqueName: \"kubernetes.io/projected/6218522b-3df9-45b0-8edb-fb935a7f4109-kube-api-access-vkwpb\") pod \"6218522b-3df9-45b0-8edb-fb935a7f4109\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.701658 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-combined-ca-bundle\") pod \"6218522b-3df9-45b0-8edb-fb935a7f4109\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.701711 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-config-data\") pod \"6218522b-3df9-45b0-8edb-fb935a7f4109\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.701749 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-nova-metadata-tls-certs\") pod \"6218522b-3df9-45b0-8edb-fb935a7f4109\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.702137 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6218522b-3df9-45b0-8edb-fb935a7f4109-logs\") pod \"6218522b-3df9-45b0-8edb-fb935a7f4109\" (UID: \"6218522b-3df9-45b0-8edb-fb935a7f4109\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.703368 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.703386 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.703395 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/897d0d90-bf27-4d5b-b136-863379656f34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.704160 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6218522b-3df9-45b0-8edb-fb935a7f4109-logs" (OuterVolumeSpecName: "logs") pod "6218522b-3df9-45b0-8edb-fb935a7f4109" (UID: "6218522b-3df9-45b0-8edb-fb935a7f4109"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.717301 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6218522b-3df9-45b0-8edb-fb935a7f4109-kube-api-access-vkwpb" (OuterVolumeSpecName: "kube-api-access-vkwpb") pod "6218522b-3df9-45b0-8edb-fb935a7f4109" (UID: "6218522b-3df9-45b0-8edb-fb935a7f4109"). InnerVolumeSpecName "kube-api-access-vkwpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.772460 4799 scope.go:117] "RemoveContainer" containerID="0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac" Nov 24 07:09:16 crc kubenswrapper[4799]: E1124 07:09:16.773632 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac\": container with ID starting with 0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac not found: ID does not exist" containerID="0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.773680 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac"} err="failed to get container status \"0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac\": rpc error: code = NotFound desc = could not find container \"0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac\": container with ID starting with 0a729eb0c4c21a8e3533e9e0e5c098dd8b521ad0fcda9988636c7b8ad73597ac not found: ID does not exist" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.773708 4799 scope.go:117] "RemoveContainer" containerID="57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.776109 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6218522b-3df9-45b0-8edb-fb935a7f4109" (UID: "6218522b-3df9-45b0-8edb-fb935a7f4109"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.794819 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-config-data" (OuterVolumeSpecName: "config-data") pod "6218522b-3df9-45b0-8edb-fb935a7f4109" (UID: "6218522b-3df9-45b0-8edb-fb935a7f4109"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.805603 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkwpb\" (UniqueName: \"kubernetes.io/projected/6218522b-3df9-45b0-8edb-fb935a7f4109-kube-api-access-vkwpb\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.805643 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.805656 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.805674 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6218522b-3df9-45b0-8edb-fb935a7f4109-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.809372 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.835616 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.843789 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell125d6-account-delete-cw4x8" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.860523 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.877097 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance00d0-account-delete-kbtpl" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.885443 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6218522b-3df9-45b0-8edb-fb935a7f4109" (UID: "6218522b-3df9-45b0-8edb-fb935a7f4109"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.908455 4799 scope.go:117] "RemoveContainer" containerID="5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.910810 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebfe77df-bbcd-412f-9703-78e23d93f410-logs\") pod \"ebfe77df-bbcd-412f-9703-78e23d93f410\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.913942 4799 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6218522b-3df9-45b0-8edb-fb935a7f4109-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.914764 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebfe77df-bbcd-412f-9703-78e23d93f410-logs" (OuterVolumeSpecName: "logs") pod "ebfe77df-bbcd-412f-9703-78e23d93f410" (UID: "ebfe77df-bbcd-412f-9703-78e23d93f410"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.936824 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.937556 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf","Type":"ContainerDied","Data":"6256ad3dda3b8c8f22ca60c97059524d69b86c6cc089604645377e1368bb932f"} Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.937630 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.941147 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderbbb1-account-delete-tjfp2" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.946666 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance00d0-account-delete-kbtpl" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.946181 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance00d0-account-delete-kbtpl" event={"ID":"ad75f933-ab36-4c9d-857d-c2cc11702dd3","Type":"ContainerDied","Data":"2ea3adf81d2d683132eb99305e52fcbd10295056e7f7ab9f970e55c50770b682"} Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.947257 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ea3adf81d2d683132eb99305e52fcbd10295056e7f7ab9f970e55c50770b682" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.949636 4799 scope.go:117] "RemoveContainer" containerID="57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e" Nov 24 07:09:16 crc kubenswrapper[4799]: E1124 07:09:16.953165 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e\": container with ID starting with 57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e not found: ID does not exist" containerID="57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.953210 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e"} err="failed to get container status \"57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e\": rpc error: code = NotFound desc = could not find container \"57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e\": container with ID starting with 57da9a98416415e796c5fb15050f5562437de8dad0e6e08c2739c444bf68280e not found: ID does not exist" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.953255 4799 scope.go:117] "RemoveContainer" containerID="5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029" Nov 24 07:09:16 crc kubenswrapper[4799]: E1124 07:09:16.954018 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029\": container with ID starting with 5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029 not found: ID does not exist" containerID="5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.954059 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029"} err="failed to get container status \"5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029\": rpc error: code = NotFound desc = could not find container \"5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029\": container with ID starting with 5eb4ecaef03ca05856ba7c5de57f18f5b499837147a52bade9e627caa6c3e029 not found: ID does not exist" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.954088 4799 scope.go:117] "RemoveContainer" containerID="183c3c900b89e4d354f72e7ee6593a3ca4a44de57d248b7f61ac4959fe51c8ef" Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.989344 4799 generic.go:334] "Generic (PLEG): container finished" podID="cb94fabc-dfc3-40cc-9791-0272924692da" containerID="dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc" exitCode=0 Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.989495 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb94fabc-dfc3-40cc-9791-0272924692da","Type":"ContainerDied","Data":"dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc"} Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.989562 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb94fabc-dfc3-40cc-9791-0272924692da","Type":"ContainerDied","Data":"50b3ca3a307bdedf42b2c4ab460d71684336723c36da0f227eb3acde7e00ec22"} Nov 24 07:09:16 crc kubenswrapper[4799]: I1124 07:09:16.989674 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.000049 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65f6685878-7bjdh" event={"ID":"897d0d90-bf27-4d5b-b136-863379656f34","Type":"ContainerDied","Data":"f1c6fcbd0f68b00e30c3cea438c780a1b0f1f06e461cd0eae3ef252321de1f0e"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.000300 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65f6685878-7bjdh" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.004344 4799 scope.go:117] "RemoveContainer" containerID="668ba3f136f6d97ba947f42f79e2823ce4845dbde12c05bc8fe71572551507f9" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.014770 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgk6w\" (UniqueName: \"kubernetes.io/projected/cb94fabc-dfc3-40cc-9791-0272924692da-kube-api-access-sgk6w\") pod \"cb94fabc-dfc3-40cc-9791-0272924692da\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.014977 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-combined-ca-bundle\") pod \"cb94fabc-dfc3-40cc-9791-0272924692da\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.015064 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb94fabc-dfc3-40cc-9791-0272924692da-etc-machine-id\") pod \"cb94fabc-dfc3-40cc-9791-0272924692da\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.015133 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc7zx\" (UniqueName: \"kubernetes.io/projected/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-kube-api-access-fc7zx\") pod \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.015231 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-internal-tls-certs\") pod \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.015309 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data\") pod \"ebfe77df-bbcd-412f-9703-78e23d93f410\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.015385 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h44k\" (UniqueName: \"kubernetes.io/projected/ebfe77df-bbcd-412f-9703-78e23d93f410-kube-api-access-8h44k\") pod \"ebfe77df-bbcd-412f-9703-78e23d93f410\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.015462 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.015550 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-httpd-run\") pod \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.015623 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-scripts\") pod \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.015713 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvfj6\" (UniqueName: \"kubernetes.io/projected/d4a27166-5918-4ee7-8569-a9f58914d7be-kube-api-access-fvfj6\") pod \"d4a27166-5918-4ee7-8569-a9f58914d7be\" (UID: \"d4a27166-5918-4ee7-8569-a9f58914d7be\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.016138 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-public-tls-certs\") pod \"ebfe77df-bbcd-412f-9703-78e23d93f410\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.016220 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data-custom\") pod \"cb94fabc-dfc3-40cc-9791-0272924692da\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.016298 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad75f933-ab36-4c9d-857d-c2cc11702dd3-operator-scripts\") pod \"ad75f933-ab36-4c9d-857d-c2cc11702dd3\" (UID: \"ad75f933-ab36-4c9d-857d-c2cc11702dd3\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.016376 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq2qg\" (UniqueName: \"kubernetes.io/projected/ad75f933-ab36-4c9d-857d-c2cc11702dd3-kube-api-access-hq2qg\") pod \"ad75f933-ab36-4c9d-857d-c2cc11702dd3\" (UID: \"ad75f933-ab36-4c9d-857d-c2cc11702dd3\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.016450 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-internal-tls-certs\") pod \"ebfe77df-bbcd-412f-9703-78e23d93f410\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.016530 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-combined-ca-bundle\") pod \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.016604 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-combined-ca-bundle\") pod \"ebfe77df-bbcd-412f-9703-78e23d93f410\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.016672 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-scripts\") pod \"cb94fabc-dfc3-40cc-9791-0272924692da\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.016740 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-config-data\") pod \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.016815 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqksl\" (UniqueName: \"kubernetes.io/projected/6fe03a76-2eb9-4728-9396-9d2dc1e743af-kube-api-access-nqksl\") pod \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.016905 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data-custom\") pod \"ebfe77df-bbcd-412f-9703-78e23d93f410\" (UID: \"ebfe77df-bbcd-412f-9703-78e23d93f410\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.017247 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-public-tls-certs\") pod \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.017376 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data\") pod \"cb94fabc-dfc3-40cc-9791-0272924692da\" (UID: \"cb94fabc-dfc3-40cc-9791-0272924692da\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.017467 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-logs\") pod \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\" (UID: \"1b14e8a8-2e19-47fd-a440-ac8d0cafecaf\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.017534 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-combined-ca-bundle\") pod \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.017619 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4a27166-5918-4ee7-8569-a9f58914d7be-operator-scripts\") pod \"d4a27166-5918-4ee7-8569-a9f58914d7be\" (UID: \"d4a27166-5918-4ee7-8569-a9f58914d7be\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.019799 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebfe77df-bbcd-412f-9703-78e23d93f410-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.021570 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4a27166-5918-4ee7-8569-a9f58914d7be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d4a27166-5918-4ee7-8569-a9f58914d7be" (UID: "d4a27166-5918-4ee7-8569-a9f58914d7be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.023781 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb94fabc-dfc3-40cc-9791-0272924692da-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cb94fabc-dfc3-40cc-9791-0272924692da" (UID: "cb94fabc-dfc3-40cc-9791-0272924692da"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.024779 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad75f933-ab36-4c9d-857d-c2cc11702dd3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad75f933-ab36-4c9d-857d-c2cc11702dd3" (UID: "ad75f933-ab36-4c9d-857d-c2cc11702dd3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.024771 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebfe77df-bbcd-412f-9703-78e23d93f410-kube-api-access-8h44k" (OuterVolumeSpecName: "kube-api-access-8h44k") pod "ebfe77df-bbcd-412f-9703-78e23d93f410" (UID: "ebfe77df-bbcd-412f-9703-78e23d93f410"). InnerVolumeSpecName "kube-api-access-8h44k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.025424 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" (UID: "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.028974 4799 generic.go:334] "Generic (PLEG): container finished" podID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerID="2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18" exitCode=0 Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.029808 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-logs" (OuterVolumeSpecName: "logs") pod "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" (UID: "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.030185 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-644b86779b-b4psv" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.031085 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644b86779b-b4psv" event={"ID":"ebfe77df-bbcd-412f-9703-78e23d93f410","Type":"ContainerDied","Data":"2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.031141 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-644b86779b-b4psv" event={"ID":"ebfe77df-bbcd-412f-9703-78e23d93f410","Type":"ContainerDied","Data":"8bb3ed080c7c54341b172b5767b6c3765958972f0eaabaa31b60e741aa82c318"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.031217 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" (UID: "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.039107 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.039233 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6fe03a76-2eb9-4728-9396-9d2dc1e743af","Type":"ContainerDied","Data":"482919ed95ee1f06b013f41ee9fc371c03d41adacb105440bc45b712fae39959"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.049215 4799 scope.go:117] "RemoveContainer" containerID="af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.051564 4799 generic.go:334] "Generic (PLEG): container finished" podID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerID="5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61" exitCode=0 Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.051712 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6218522b-3df9-45b0-8edb-fb935a7f4109","Type":"ContainerDied","Data":"5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.051762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6218522b-3df9-45b0-8edb-fb935a7f4109","Type":"ContainerDied","Data":"a0fadf5e8fdde0305f30ee9718c93de80bc59c9bf6e5f5cb0dc63c81954d2054"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.052045 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.058354 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-kube-api-access-fc7zx" (OuterVolumeSpecName: "kube-api-access-fc7zx") pod "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" (UID: "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf"). InnerVolumeSpecName "kube-api-access-fc7zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.066572 4799 generic.go:334] "Generic (PLEG): container finished" podID="168550d3-6597-41ff-adf4-d20f190c7a02" containerID="192c554265f38b0137575a7bb2ad8ef843ac317c21ba65066896ec0465f061f3" exitCode=0 Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.066648 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"168550d3-6597-41ff-adf4-d20f190c7a02","Type":"ContainerDied","Data":"192c554265f38b0137575a7bb2ad8ef843ac317c21ba65066896ec0465f061f3"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.067619 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad75f933-ab36-4c9d-857d-c2cc11702dd3-kube-api-access-hq2qg" (OuterVolumeSpecName: "kube-api-access-hq2qg") pod "ad75f933-ab36-4c9d-857d-c2cc11702dd3" (UID: "ad75f933-ab36-4c9d-857d-c2cc11702dd3"). InnerVolumeSpecName "kube-api-access-hq2qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.068724 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fe03a76-2eb9-4728-9396-9d2dc1e743af-kube-api-access-nqksl" (OuterVolumeSpecName: "kube-api-access-nqksl") pod "6fe03a76-2eb9-4728-9396-9d2dc1e743af" (UID: "6fe03a76-2eb9-4728-9396-9d2dc1e743af"). InnerVolumeSpecName "kube-api-access-nqksl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.068631 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-scripts" (OuterVolumeSpecName: "scripts") pod "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" (UID: "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.070742 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb94fabc-dfc3-40cc-9791-0272924692da-kube-api-access-sgk6w" (OuterVolumeSpecName: "kube-api-access-sgk6w") pod "cb94fabc-dfc3-40cc-9791-0272924692da" (UID: "cb94fabc-dfc3-40cc-9791-0272924692da"). InnerVolumeSpecName "kube-api-access-sgk6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.070755 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-scripts" (OuterVolumeSpecName: "scripts") pod "cb94fabc-dfc3-40cc-9791-0272924692da" (UID: "cb94fabc-dfc3-40cc-9791-0272924692da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.072412 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cb94fabc-dfc3-40cc-9791-0272924692da" (UID: "cb94fabc-dfc3-40cc-9791-0272924692da"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.072626 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4a27166-5918-4ee7-8569-a9f58914d7be-kube-api-access-fvfj6" (OuterVolumeSpecName: "kube-api-access-fvfj6") pod "d4a27166-5918-4ee7-8569-a9f58914d7be" (UID: "d4a27166-5918-4ee7-8569-a9f58914d7be"). InnerVolumeSpecName "kube-api-access-fvfj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.083014 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ebfe77df-bbcd-412f-9703-78e23d93f410" (UID: "ebfe77df-bbcd-412f-9703-78e23d93f410"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.084743 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e9a036d8-ce56-4654-a782-15cd18730378","Type":"ContainerDied","Data":"0b7a475aa5bac703f2bfb4027d1c6068f6d36bf5c155cd26fa56d69d8fe7c273"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.085092 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.104673 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderbbb1-account-delete-tjfp2" event={"ID":"d9064d19-6806-4e78-beb4-cda924fe3010","Type":"ContainerDied","Data":"eff18abe0edf7afb956eb450e236f49b9a829daf1324769ee95e5fc25e5bf28c"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.104720 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eff18abe0edf7afb956eb450e236f49b9a829daf1324769ee95e5fc25e5bf28c" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.104777 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderbbb1-account-delete-tjfp2" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.106712 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.111296 4799 generic.go:334] "Generic (PLEG): container finished" podID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerID="dc5fdc4d831816407f510ddb2a4fc3bcb5bbc530983dfb627e98b50af968d2bd" exitCode=0 Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.111333 4799 generic.go:334] "Generic (PLEG): container finished" podID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerID="4b1017dff22faa9e7cf62359f3de7ceea284aa7bf591e9d872e15ff649d22d96" exitCode=0 Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.111380 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f97d9a56-cb78-4da9-b832-0028b6e45895","Type":"ContainerDied","Data":"dc5fdc4d831816407f510ddb2a4fc3bcb5bbc530983dfb627e98b50af968d2bd"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.111408 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f97d9a56-cb78-4da9-b832-0028b6e45895","Type":"ContainerDied","Data":"4b1017dff22faa9e7cf62359f3de7ceea284aa7bf591e9d872e15ff649d22d96"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.114095 4799 generic.go:334] "Generic (PLEG): container finished" podID="b0361baf-d0c3-430f-87e3-c011f869a2b3" containerID="4bcf19e3b710fbf877a92e707435d95962e55b4801e15a2e5b43f689d28b8bba" exitCode=0 Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.114154 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b0361baf-d0c3-430f-87e3-c011f869a2b3","Type":"ContainerDied","Data":"4bcf19e3b710fbf877a92e707435d95962e55b4801e15a2e5b43f689d28b8bba"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.114215 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.114608 4799 scope.go:117] "RemoveContainer" containerID="dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.120526 4799 generic.go:334] "Generic (PLEG): container finished" podID="d4a27166-5918-4ee7-8569-a9f58914d7be" containerID="36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330" exitCode=1 Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.120611 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell125d6-account-delete-cw4x8" event={"ID":"d4a27166-5918-4ee7-8569-a9f58914d7be","Type":"ContainerDied","Data":"36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.120644 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell125d6-account-delete-cw4x8" event={"ID":"d4a27166-5918-4ee7-8569-a9f58914d7be","Type":"ContainerDied","Data":"51e4e0483f26c31669ce075027014e0a4faa36f8f873839a0c1c1a89d908f12d"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.120702 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell125d6-account-delete-cw4x8" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.121176 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9064d19-6806-4e78-beb4-cda924fe3010-operator-scripts\") pod \"d9064d19-6806-4e78-beb4-cda924fe3010\" (UID: \"d9064d19-6806-4e78-beb4-cda924fe3010\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.121237 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-internal-tls-certs\") pod \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.121412 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-kolla-config\") pod \"b0361baf-d0c3-430f-87e3-c011f869a2b3\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.121449 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl9kp\" (UniqueName: \"kubernetes.io/projected/b0361baf-d0c3-430f-87e3-c011f869a2b3-kube-api-access-jl9kp\") pod \"b0361baf-d0c3-430f-87e3-c011f869a2b3\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.121487 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-config-data\") pod \"b0361baf-d0c3-430f-87e3-c011f869a2b3\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.121531 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-combined-ca-bundle\") pod \"b0361baf-d0c3-430f-87e3-c011f869a2b3\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.121549 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-memcached-tls-certs\") pod \"b0361baf-d0c3-430f-87e3-c011f869a2b3\" (UID: \"b0361baf-d0c3-430f-87e3-c011f869a2b3\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.121581 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-config-data\") pod \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.121612 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx9gx\" (UniqueName: \"kubernetes.io/projected/d9064d19-6806-4e78-beb4-cda924fe3010-kube-api-access-wx9gx\") pod \"d9064d19-6806-4e78-beb4-cda924fe3010\" (UID: \"d9064d19-6806-4e78-beb4-cda924fe3010\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.121642 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fe03a76-2eb9-4728-9396-9d2dc1e743af-logs\") pod \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\" (UID: \"6fe03a76-2eb9-4728-9396-9d2dc1e743af\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.123786 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.123829 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.123843 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvfj6\" (UniqueName: \"kubernetes.io/projected/d4a27166-5918-4ee7-8569-a9f58914d7be-kube-api-access-fvfj6\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.123927 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.123941 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad75f933-ab36-4c9d-857d-c2cc11702dd3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.123978 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq2qg\" (UniqueName: \"kubernetes.io/projected/ad75f933-ab36-4c9d-857d-c2cc11702dd3-kube-api-access-hq2qg\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.123993 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.124002 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqksl\" (UniqueName: \"kubernetes.io/projected/6fe03a76-2eb9-4728-9396-9d2dc1e743af-kube-api-access-nqksl\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.124014 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.124023 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.124031 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4a27166-5918-4ee7-8569-a9f58914d7be-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.124041 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgk6w\" (UniqueName: \"kubernetes.io/projected/cb94fabc-dfc3-40cc-9791-0272924692da-kube-api-access-sgk6w\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.124050 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb94fabc-dfc3-40cc-9791-0272924692da-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.124059 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc7zx\" (UniqueName: \"kubernetes.io/projected/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-kube-api-access-fc7zx\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.124070 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h44k\" (UniqueName: \"kubernetes.io/projected/ebfe77df-bbcd-412f-9703-78e23d93f410-kube-api-access-8h44k\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.124093 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.135967 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9064d19-6806-4e78-beb4-cda924fe3010-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d9064d19-6806-4e78-beb4-cda924fe3010" (UID: "d9064d19-6806-4e78-beb4-cda924fe3010"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.136754 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fe03a76-2eb9-4728-9396-9d2dc1e743af-logs" (OuterVolumeSpecName: "logs") pod "6fe03a76-2eb9-4728-9396-9d2dc1e743af" (UID: "6fe03a76-2eb9-4728-9396-9d2dc1e743af"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.136986 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "b0361baf-d0c3-430f-87e3-c011f869a2b3" (UID: "b0361baf-d0c3-430f-87e3-c011f869a2b3"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.137638 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-config-data" (OuterVolumeSpecName: "config-data") pod "b0361baf-d0c3-430f-87e3-c011f869a2b3" (UID: "b0361baf-d0c3-430f-87e3-c011f869a2b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.139673 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.141985 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.142043 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3cd44457-1418-4788-a029-82c54c917997","Type":"ContainerDied","Data":"63de444daa10203a26673c85620a0d616f43b8136c3953d59df36db67362f926"} Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.142202 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.142345 4799 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican44f2-account-delete-2hm9q" secret="" err="secret \"galera-openstack-dockercfg-hptzz\" not found" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.143392 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell000c8-account-delete-n5mhs" podUID="87439e47-6c84-4467-a49b-03ed6e86a9eb" containerName="mariadb-account-delete" containerID="cri-o://44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6" gracePeriod=30 Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.143544 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement2704-account-delete-ck9t2" podUID="b6f22313-6d9a-4b35-b892-695903d354b1" containerName="mariadb-account-delete" containerID="cri-o://fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61" gracePeriod=30 Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.158855 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9064d19-6806-4e78-beb4-cda924fe3010-kube-api-access-wx9gx" (OuterVolumeSpecName: "kube-api-access-wx9gx") pod "d9064d19-6806-4e78-beb4-cda924fe3010" (UID: "d9064d19-6806-4e78-beb4-cda924fe3010"). InnerVolumeSpecName "kube-api-access-wx9gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.187652 4799 scope.go:117] "RemoveContainer" containerID="af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.187774 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0361baf-d0c3-430f-87e3-c011f869a2b3-kube-api-access-jl9kp" (OuterVolumeSpecName: "kube-api-access-jl9kp") pod "b0361baf-d0c3-430f-87e3-c011f869a2b3" (UID: "b0361baf-d0c3-430f-87e3-c011f869a2b3"). InnerVolumeSpecName "kube-api-access-jl9kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.189268 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5\": container with ID starting with af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5 not found: ID does not exist" containerID="af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.189308 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5"} err="failed to get container status \"af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5\": rpc error: code = NotFound desc = could not find container \"af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5\": container with ID starting with af75b91f5fa58c28caac8ef9ff3a0b716a9f1b230f7fc6cd8303c79e16e4a5f5 not found: ID does not exist" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.189337 4799 scope.go:117] "RemoveContainer" containerID="dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc" Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.191118 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc\": container with ID starting with dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc not found: ID does not exist" containerID="dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.191144 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc"} err="failed to get container status \"dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc\": rpc error: code = NotFound desc = could not find container \"dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc\": container with ID starting with dc3856e5432f85617ca02fb908cd8e624a2cb1c9a73894553217061453243acc not found: ID does not exist" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.191159 4799 scope.go:117] "RemoveContainer" containerID="62f862c6ff3823a347d374cd9375930cec2ce5b254235146abb2a0f3a2a1a61c" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.225243 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-combined-ca-bundle\") pod \"168550d3-6597-41ff-adf4-d20f190c7a02\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.225321 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-config-data\") pod \"168550d3-6597-41ff-adf4-d20f190c7a02\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.225416 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65t49\" (UniqueName: \"kubernetes.io/projected/168550d3-6597-41ff-adf4-d20f190c7a02-kube-api-access-65t49\") pod \"168550d3-6597-41ff-adf4-d20f190c7a02\" (UID: \"168550d3-6597-41ff-adf4-d20f190c7a02\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.225785 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9064d19-6806-4e78-beb4-cda924fe3010-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.225801 4799 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.225813 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl9kp\" (UniqueName: \"kubernetes.io/projected/b0361baf-d0c3-430f-87e3-c011f869a2b3-kube-api-access-jl9kp\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.225826 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0361baf-d0c3-430f-87e3-c011f869a2b3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.225839 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx9gx\" (UniqueName: \"kubernetes.io/projected/d9064d19-6806-4e78-beb4-cda924fe3010-kube-api-access-wx9gx\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.225870 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fe03a76-2eb9-4728-9396-9d2dc1e743af-logs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.226292 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebfe77df-bbcd-412f-9703-78e23d93f410" (UID: "ebfe77df-bbcd-412f-9703-78e23d93f410"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.230481 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-config-data" (OuterVolumeSpecName: "config-data") pod "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" (UID: "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.240155 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ebfe77df-bbcd-412f-9703-78e23d93f410" (UID: "ebfe77df-bbcd-412f-9703-78e23d93f410"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.241374 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6fe03a76-2eb9-4728-9396-9d2dc1e743af" (UID: "6fe03a76-2eb9-4728-9396-9d2dc1e743af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.262448 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/168550d3-6597-41ff-adf4-d20f190c7a02-kube-api-access-65t49" (OuterVolumeSpecName: "kube-api-access-65t49") pod "168550d3-6597-41ff-adf4-d20f190c7a02" (UID: "168550d3-6597-41ff-adf4-d20f190c7a02"). InnerVolumeSpecName "kube-api-access-65t49". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.262617 4799 scope.go:117] "RemoveContainer" containerID="702c77d5a08ac68b078c369149870d0f8a51352018417fd5ffd7488783fd889b" Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.265679 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.267327 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.275695 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.275737 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerName="ovn-northd" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.279790 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" (UID: "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.295314 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.301563 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0361baf-d0c3-430f-87e3-c011f869a2b3" (UID: "b0361baf-d0c3-430f-87e3-c011f869a2b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.307763 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-config-data" (OuterVolumeSpecName: "config-data") pod "6fe03a76-2eb9-4728-9396-9d2dc1e743af" (UID: "6fe03a76-2eb9-4728-9396-9d2dc1e743af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.311081 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-65f6685878-7bjdh"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.321949 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data" (OuterVolumeSpecName: "config-data") pod "ebfe77df-bbcd-412f-9703-78e23d93f410" (UID: "ebfe77df-bbcd-412f-9703-78e23d93f410"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.324997 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-65f6685878-7bjdh"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.327612 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.327663 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.327677 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65t49\" (UniqueName: \"kubernetes.io/projected/168550d3-6597-41ff-adf4-d20f190c7a02-kube-api-access-65t49\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.327693 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.327705 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.327715 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.327726 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.327735 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.327745 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.327756 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.337588 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.347714 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb94fabc-dfc3-40cc-9791-0272924692da" (UID: "cb94fabc-dfc3-40cc-9791-0272924692da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.352401 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.352440 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ebfe77df-bbcd-412f-9703-78e23d93f410" (UID: "ebfe77df-bbcd-412f-9703-78e23d93f410"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.360955 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" (UID: "1b14e8a8-2e19-47fd-a440-ac8d0cafecaf"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.369385 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell125d6-account-delete-cw4x8"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.371417 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "168550d3-6597-41ff-adf4-d20f190c7a02" (UID: "168550d3-6597-41ff-adf4-d20f190c7a02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.374516 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "b0361baf-d0c3-430f-87e3-c011f869a2b3" (UID: "b0361baf-d0c3-430f-87e3-c011f869a2b3"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.376742 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6fe03a76-2eb9-4728-9396-9d2dc1e743af" (UID: "6fe03a76-2eb9-4728-9396-9d2dc1e743af"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.378677 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell125d6-account-delete-cw4x8"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.385484 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.385640 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6fe03a76-2eb9-4728-9396-9d2dc1e743af" (UID: "6fe03a76-2eb9-4728-9396-9d2dc1e743af"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.391543 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.391596 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-config-data" (OuterVolumeSpecName: "config-data") pod "168550d3-6597-41ff-adf4-d20f190c7a02" (UID: "168550d3-6597-41ff-adf4-d20f190c7a02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.403047 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.407960 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.412940 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.418802 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.427957 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data" (OuterVolumeSpecName: "config-data") pod "cb94fabc-dfc3-40cc-9791-0272924692da" (UID: "cb94fabc-dfc3-40cc-9791-0272924692da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.428945 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebfe77df-bbcd-412f-9703-78e23d93f410-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.429281 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.429367 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.429430 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.429488 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/168550d3-6597-41ff-adf4-d20f190c7a02-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.429573 4799 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0361baf-d0c3-430f-87e3-c011f869a2b3-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.429638 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb94fabc-dfc3-40cc-9791-0272924692da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.429698 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.429759 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fe03a76-2eb9-4728-9396-9d2dc1e743af-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.529564 4799 scope.go:117] "RemoveContainer" containerID="2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18" Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.531003 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.531388 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts podName:87439e47-6c84-4467-a49b-03ed6e86a9eb nodeName:}" failed. No retries permitted until 2025-11-24 07:09:19.531373828 +0000 UTC m=+1305.187356302 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts") pod "novacell000c8-account-delete-n5mhs" (UID: "87439e47-6c84-4467-a49b-03ed6e86a9eb") : configmap "openstack-scripts" not found Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.531192 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.531870 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts podName:d13b66c3-1e31-418e-a0ff-0d1bf7a5980a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:19.531834101 +0000 UTC m=+1305.187816655 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts") pod "barbican44f2-account-delete-2hm9q" (UID: "d13b66c3-1e31-418e-a0ff-0d1bf7a5980a") : configmap "openstack-scripts" not found Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.553999 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.568211 4799 scope.go:117] "RemoveContainer" containerID="c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.569498 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.585357 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.595829 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.608252 4799 scope.go:117] "RemoveContainer" containerID="2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18" Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.608747 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18\": container with ID starting with 2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18 not found: ID does not exist" containerID="2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.608825 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18"} err="failed to get container status \"2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18\": rpc error: code = NotFound desc = could not find container \"2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18\": container with ID starting with 2d70b91fcadd84f92bb789beb1506c335283981a8f2480c624793df46a7b8d18 not found: ID does not exist" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.608883 4799 scope.go:117] "RemoveContainer" containerID="c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af" Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.609340 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af\": container with ID starting with c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af not found: ID does not exist" containerID="c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.609376 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af"} err="failed to get container status \"c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af\": rpc error: code = NotFound desc = could not find container \"c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af\": container with ID starting with c97cbee439a46638bf99832f7bc6417528eebd687f118d797e4fad9fa73626af not found: ID does not exist" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.609396 4799 scope.go:117] "RemoveContainer" containerID="f6b04750444de5a3182982a1b53fc81a035e9a9f2eb6df3d21f831648f26fc49" Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.631955 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.632096 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts podName:b6f22313-6d9a-4b35-b892-695903d354b1 nodeName:}" failed. No retries permitted until 2025-11-24 07:09:19.632064585 +0000 UTC m=+1305.288047059 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts") pod "placement2704-account-delete-ck9t2" (UID: "b6f22313-6d9a-4b35-b892-695903d354b1") : configmap "openstack-scripts" not found Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.632234 4799 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.632327 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data podName:96acf88b-3fe8-46ac-8393-37f720632dd6 nodeName:}" failed. No retries permitted until 2025-11-24 07:09:25.632306952 +0000 UTC m=+1311.288289606 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data") pod "rabbitmq-server-0" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6") : configmap "rabbitmq-config-data" not found Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.641778 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04ddaef5-8849-4fa9-a0b9-f5e009dff420" path="/var/lib/kubelet/pods/04ddaef5-8849-4fa9-a0b9-f5e009dff420/volumes" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.642397 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" path="/var/lib/kubelet/pods/1b14e8a8-2e19-47fd-a440-ac8d0cafecaf/volumes" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.642980 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd44457-1418-4788-a029-82c54c917997" path="/var/lib/kubelet/pods/3cd44457-1418-4788-a029-82c54c917997/volumes" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.648688 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dd8bba3-489f-442a-97a4-47d2ebffdc9b" path="/var/lib/kubelet/pods/3dd8bba3-489f-442a-97a4-47d2ebffdc9b/volumes" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.650998 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" path="/var/lib/kubelet/pods/6218522b-3df9-45b0-8edb-fb935a7f4109/volumes" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.651632 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7130f73f-5578-4310-96f3-db076a954990" path="/var/lib/kubelet/pods/7130f73f-5578-4310-96f3-db076a954990/volumes" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.654391 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="897d0d90-bf27-4d5b-b136-863379656f34" path="/var/lib/kubelet/pods/897d0d90-bf27-4d5b-b136-863379656f34/volumes" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.656221 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0361baf-d0c3-430f-87e3-c011f869a2b3" path="/var/lib/kubelet/pods/b0361baf-d0c3-430f-87e3-c011f869a2b3/volumes" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.658086 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4a27166-5918-4ee7-8569-a9f58914d7be" path="/var/lib/kubelet/pods/d4a27166-5918-4ee7-8569-a9f58914d7be/volumes" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.659225 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9a036d8-ce56-4654-a782-15cd18730378" path="/var/lib/kubelet/pods/e9a036d8-ce56-4654-a782-15cd18730378/volumes" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.661705 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f643414c-72f0-4e7c-b804-20a56f3207ab" path="/var/lib/kubelet/pods/f643414c-72f0-4e7c-b804-20a56f3207ab/volumes" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.717307 4799 scope.go:117] "RemoveContainer" containerID="cbf304821538247ce0d3cfcbea062c7ba81f6603d05b995bc8ebc90500583b62" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.717375 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron2d2f-account-delete-5c2sd" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.736068 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-operator-scripts\") pod \"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee\" (UID: \"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.736103 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zn9g\" (UniqueName: \"kubernetes.io/projected/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-kube-api-access-8zn9g\") pod \"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee\" (UID: \"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.737437 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3f86cb1-3f0e-4da8-82dd-e00522d4ceee" (UID: "d3f86cb1-3f0e-4da8-82dd-e00522d4ceee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.748097 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi722e-account-delete-th6f8" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.754791 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-kube-api-access-8zn9g" (OuterVolumeSpecName: "kube-api-access-8zn9g") pod "d3f86cb1-3f0e-4da8-82dd-e00522d4ceee" (UID: "d3f86cb1-3f0e-4da8-82dd-e00522d4ceee"). InnerVolumeSpecName "kube-api-access-8zn9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.764256 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.771089 4799 scope.go:117] "RemoveContainer" containerID="5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.796413 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.820918 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.832748 4799 scope.go:117] "RemoveContainer" containerID="c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.832929 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.837556 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63757cd4-2ed0-4423-b466-4dfac76e3ac4-operator-scripts\") pod \"63757cd4-2ed0-4423-b466-4dfac76e3ac4\" (UID: \"63757cd4-2ed0-4423-b466-4dfac76e3ac4\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.837644 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5sfx\" (UniqueName: \"kubernetes.io/projected/63757cd4-2ed0-4423-b466-4dfac76e3ac4-kube-api-access-j5sfx\") pod \"63757cd4-2ed0-4423-b466-4dfac76e3ac4\" (UID: \"63757cd4-2ed0-4423-b466-4dfac76e3ac4\") " Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.837966 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.837991 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zn9g\" (UniqueName: \"kubernetes.io/projected/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee-kube-api-access-8zn9g\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.839116 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63757cd4-2ed0-4423-b466-4dfac76e3ac4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "63757cd4-2ed0-4423-b466-4dfac76e3ac4" (UID: "63757cd4-2ed0-4423-b466-4dfac76e3ac4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.843075 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-644b86779b-b4psv"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.843817 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63757cd4-2ed0-4423-b466-4dfac76e3ac4-kube-api-access-j5sfx" (OuterVolumeSpecName: "kube-api-access-j5sfx") pod "63757cd4-2ed0-4423-b466-4dfac76e3ac4" (UID: "63757cd4-2ed0-4423-b466-4dfac76e3ac4"). InnerVolumeSpecName "kube-api-access-j5sfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.854937 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-644b86779b-b4psv"] Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.856566 4799 scope.go:117] "RemoveContainer" containerID="5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61" Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.857043 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61\": container with ID starting with 5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61 not found: ID does not exist" containerID="5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.857095 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61"} err="failed to get container status \"5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61\": rpc error: code = NotFound desc = could not find container \"5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61\": container with ID starting with 5dc8145b5f7d2b22505a05f5bffe6c18ca208bf1b5e05f6b69aaf61600f7dd61 not found: ID does not exist" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.857127 4799 scope.go:117] "RemoveContainer" containerID="c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547" Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.857524 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547\": container with ID starting with c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547 not found: ID does not exist" containerID="c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.857572 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547"} err="failed to get container status \"c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547\": rpc error: code = NotFound desc = could not find container \"c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547\": container with ID starting with c1a2fd9db3c29c8fb328f819c44cc47e8035dab5efd5ca0394d7adf7d3376547 not found: ID does not exist" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.857603 4799 scope.go:117] "RemoveContainer" containerID="e97f64b93e8c87fb3e9989776d9be9cfbdd54d15ade6362a154f53220314a180" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.882546 4799 scope.go:117] "RemoveContainer" containerID="bb47a5f8acf503e4173095e108c8eaae5b42159d81d170336e69837e3fecd4b9" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.903089 4799 scope.go:117] "RemoveContainer" containerID="4bcf19e3b710fbf877a92e707435d95962e55b4801e15a2e5b43f689d28b8bba" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.937685 4799 scope.go:117] "RemoveContainer" containerID="36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.940404 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63757cd4-2ed0-4423-b466-4dfac76e3ac4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.940474 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5sfx\" (UniqueName: \"kubernetes.io/projected/63757cd4-2ed0-4423-b466-4dfac76e3ac4-kube-api-access-j5sfx\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.970115 4799 scope.go:117] "RemoveContainer" containerID="36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330" Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.970738 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330\": container with ID starting with 36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330 not found: ID does not exist" containerID="36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.970766 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330"} err="failed to get container status \"36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330\": rpc error: code = NotFound desc = could not find container \"36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330\": container with ID starting with 36f2bbea0880435b82407831ea911beec1c1add3f417c65caa4bcef4c92ee330 not found: ID does not exist" Nov 24 07:09:17 crc kubenswrapper[4799]: I1124 07:09:17.970788 4799 scope.go:117] "RemoveContainer" containerID="e7c302a760dcca397dd206e8bf79d8516ca32ba49a1ef37a43e67e1893d9366d" Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.987182 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 07:09:17 crc kubenswrapper[4799]: E1124 07:09:17.991611 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 07:09:18 crc kubenswrapper[4799]: E1124 07:09:18.004333 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 07:09:18 crc kubenswrapper[4799]: E1124 07:09:18.004423 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="66ae5dfe-6c96-4d2d-99cf-b792b139d1a1" containerName="nova-cell0-conductor-conductor" Nov 24 07:09:18 crc kubenswrapper[4799]: E1124 07:09:18.041539 4799 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 07:09:18 crc kubenswrapper[4799]: E1124 07:09:18.041595 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data podName:366b7548-2fca-4623-a45c-c1e0367ce93a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:26.041581956 +0000 UTC m=+1311.697564430 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data") pod "rabbitmq-cell1-server-0" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a") : configmap "rabbitmq-cell1-config-data" not found Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.044241 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.145230 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-generated\") pod \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.145295 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkvnz\" (UniqueName: \"kubernetes.io/projected/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kube-api-access-dkvnz\") pod \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.145338 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-combined-ca-bundle\") pod \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.145405 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kolla-config\") pod \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.145446 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-galera-tls-certs\") pod \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.146457 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.146508 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-operator-scripts\") pod \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.146525 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "997ffb9e-cc6a-4fca-91d9-eb17d711b135" (UID: "997ffb9e-cc6a-4fca-91d9-eb17d711b135"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.146536 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-default\") pod \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\" (UID: \"997ffb9e-cc6a-4fca-91d9-eb17d711b135\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.146540 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "997ffb9e-cc6a-4fca-91d9-eb17d711b135" (UID: "997ffb9e-cc6a-4fca-91d9-eb17d711b135"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.147214 4799 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.147240 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.147772 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "997ffb9e-cc6a-4fca-91d9-eb17d711b135" (UID: "997ffb9e-cc6a-4fca-91d9-eb17d711b135"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.148103 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "997ffb9e-cc6a-4fca-91d9-eb17d711b135" (UID: "997ffb9e-cc6a-4fca-91d9-eb17d711b135"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.152823 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"168550d3-6597-41ff-adf4-d20f190c7a02","Type":"ContainerDied","Data":"b9bd0d67bff6feeb67b72850f03b07e419a11d9f67848b25fdefe0da91074088"} Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.152972 4799 scope.go:117] "RemoveContainer" containerID="192c554265f38b0137575a7bb2ad8ef843ac317c21ba65066896ec0465f061f3" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.153068 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kube-api-access-dkvnz" (OuterVolumeSpecName: "kube-api-access-dkvnz") pod "997ffb9e-cc6a-4fca-91d9-eb17d711b135" (UID: "997ffb9e-cc6a-4fca-91d9-eb17d711b135"). InnerVolumeSpecName "kube-api-access-dkvnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.153096 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.158347 4799 generic.go:334] "Generic (PLEG): container finished" podID="366b7548-2fca-4623-a45c-c1e0367ce93a" containerID="54abd42b7590aea7f2db6c595fca4f645386857fcbdb38711dfb864deb083510" exitCode=0 Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.158446 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"366b7548-2fca-4623-a45c-c1e0367ce93a","Type":"ContainerDied","Data":"54abd42b7590aea7f2db6c595fca4f645386857fcbdb38711dfb864deb083510"} Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.163313 4799 generic.go:334] "Generic (PLEG): container finished" podID="997ffb9e-cc6a-4fca-91d9-eb17d711b135" containerID="63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174" exitCode=0 Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.163377 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.163376 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"997ffb9e-cc6a-4fca-91d9-eb17d711b135","Type":"ContainerDied","Data":"63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174"} Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.164000 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"997ffb9e-cc6a-4fca-91d9-eb17d711b135","Type":"ContainerDied","Data":"da9ce81af35a9eaee55e0fbdda1e265b5056838f3b48fe78debfd2804a6c6596"} Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.176155 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "mysql-db") pod "997ffb9e-cc6a-4fca-91d9-eb17d711b135" (UID: "997ffb9e-cc6a-4fca-91d9-eb17d711b135"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.201327 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.206873 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "997ffb9e-cc6a-4fca-91d9-eb17d711b135" (UID: "997ffb9e-cc6a-4fca-91d9-eb17d711b135"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.208241 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron2d2f-account-delete-5c2sd" event={"ID":"d3f86cb1-3f0e-4da8-82dd-e00522d4ceee","Type":"ContainerDied","Data":"3e6f01b2ffc127b14096eb3eb3ce0f7eb3f9e8f18f4b5d85f347feda67faa3e4"} Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.208290 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e6f01b2ffc127b14096eb3eb3ce0f7eb3f9e8f18f4b5d85f347feda67faa3e4" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.208371 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron2d2f-account-delete-5c2sd" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.213944 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.222996 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "997ffb9e-cc6a-4fca-91d9-eb17d711b135" (UID: "997ffb9e-cc6a-4fca-91d9-eb17d711b135"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.232262 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi722e-account-delete-th6f8" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.232271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi722e-account-delete-th6f8" event={"ID":"63757cd4-2ed0-4423-b466-4dfac76e3ac4","Type":"ContainerDied","Data":"0de054a9941a3f83dadfba70e8e2636c51690ca41971fa85674b31ef65e3d353"} Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.232300 4799 scope.go:117] "RemoveContainer" containerID="63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.253639 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkvnz\" (UniqueName: \"kubernetes.io/projected/997ffb9e-cc6a-4fca-91d9-eb17d711b135-kube-api-access-dkvnz\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.254512 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.254531 4799 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/997ffb9e-cc6a-4fca-91d9-eb17d711b135-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.254553 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.254563 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.254596 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/997ffb9e-cc6a-4fca-91d9-eb17d711b135-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.287363 4799 scope.go:117] "RemoveContainer" containerID="664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.287572 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi722e-account-delete-th6f8"] Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.292268 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi722e-account-delete-th6f8"] Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.310956 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.342816 4799 scope.go:117] "RemoveContainer" containerID="63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174" Nov 24 07:09:18 crc kubenswrapper[4799]: E1124 07:09:18.345587 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174\": container with ID starting with 63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174 not found: ID does not exist" containerID="63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.345663 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174"} err="failed to get container status \"63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174\": rpc error: code = NotFound desc = could not find container \"63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174\": container with ID starting with 63d57ae956ce11c6bc74b0218ba3aebc6edc088eb80e324e5abd74d6c2c03174 not found: ID does not exist" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.345714 4799 scope.go:117] "RemoveContainer" containerID="664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2" Nov 24 07:09:18 crc kubenswrapper[4799]: E1124 07:09:18.346382 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2\": container with ID starting with 664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2 not found: ID does not exist" containerID="664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.346418 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2"} err="failed to get container status \"664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2\": rpc error: code = NotFound desc = could not find container \"664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2\": container with ID starting with 664446e580e710f5802b75734241177e105b7f51ccc41030ee70a09712337ff2 not found: ID does not exist" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.346445 4799 scope.go:117] "RemoveContainer" containerID="ecec47ab274347df90c4772ded8aa7a5894d4529edf5f63f16082a56ca0c1da3" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.356388 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.507454 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.514749 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.542092 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.661968 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-plugins-conf\") pod \"366b7548-2fca-4623-a45c-c1e0367ce93a\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662082 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/366b7548-2fca-4623-a45c-c1e0367ce93a-pod-info\") pod \"366b7548-2fca-4623-a45c-c1e0367ce93a\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662149 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data\") pod \"366b7548-2fca-4623-a45c-c1e0367ce93a\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662183 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-erlang-cookie\") pod \"366b7548-2fca-4623-a45c-c1e0367ce93a\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662247 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-confd\") pod \"366b7548-2fca-4623-a45c-c1e0367ce93a\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662269 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"366b7548-2fca-4623-a45c-c1e0367ce93a\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662363 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/366b7548-2fca-4623-a45c-c1e0367ce93a-erlang-cookie-secret\") pod \"366b7548-2fca-4623-a45c-c1e0367ce93a\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662393 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwkm6\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-kube-api-access-dwkm6\") pod \"366b7548-2fca-4623-a45c-c1e0367ce93a\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662431 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-server-conf\") pod \"366b7548-2fca-4623-a45c-c1e0367ce93a\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662449 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-plugins\") pod \"366b7548-2fca-4623-a45c-c1e0367ce93a\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662509 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-tls\") pod \"366b7548-2fca-4623-a45c-c1e0367ce93a\" (UID: \"366b7548-2fca-4623-a45c-c1e0367ce93a\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662608 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "366b7548-2fca-4623-a45c-c1e0367ce93a" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.662920 4799 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.665429 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "366b7548-2fca-4623-a45c-c1e0367ce93a" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.665677 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "366b7548-2fca-4623-a45c-c1e0367ce93a" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.666030 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.671866 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366b7548-2fca-4623-a45c-c1e0367ce93a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "366b7548-2fca-4623-a45c-c1e0367ce93a" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.672004 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "366b7548-2fca-4623-a45c-c1e0367ce93a" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.672227 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-kube-api-access-dwkm6" (OuterVolumeSpecName: "kube-api-access-dwkm6") pod "366b7548-2fca-4623-a45c-c1e0367ce93a" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a"). InnerVolumeSpecName "kube-api-access-dwkm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.672281 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "366b7548-2fca-4623-a45c-c1e0367ce93a" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.674316 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/366b7548-2fca-4623-a45c-c1e0367ce93a-pod-info" (OuterVolumeSpecName: "pod-info") pod "366b7548-2fca-4623-a45c-c1e0367ce93a" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.701547 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data" (OuterVolumeSpecName: "config-data") pod "366b7548-2fca-4623-a45c-c1e0367ce93a" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.751901 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-server-conf" (OuterVolumeSpecName: "server-conf") pod "366b7548-2fca-4623-a45c-c1e0367ce93a" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.764140 4799 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.764377 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.765335 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.765382 4799 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/366b7548-2fca-4623-a45c-c1e0367ce93a-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.765399 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/366b7548-2fca-4623-a45c-c1e0367ce93a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.765423 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.765454 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.766147 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6aa59428-ee60-4434-9245-3ff8fa55200e/ovn-northd/0.log" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.766294 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.766537 4799 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/366b7548-2fca-4623-a45c-c1e0367ce93a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.766587 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwkm6\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-kube-api-access-dwkm6\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.787675 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.787874 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "366b7548-2fca-4623-a45c-c1e0367ce93a" (UID: "366b7548-2fca-4623-a45c-c1e0367ce93a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.868434 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-confd\") pod \"96acf88b-3fe8-46ac-8393-37f720632dd6\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.868491 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh8mr\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-kube-api-access-wh8mr\") pod \"96acf88b-3fe8-46ac-8393-37f720632dd6\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.868645 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-plugins-conf\") pod \"96acf88b-3fe8-46ac-8393-37f720632dd6\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869262 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-scripts\") pod \"6aa59428-ee60-4434-9245-3ff8fa55200e\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869318 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb55z\" (UniqueName: \"kubernetes.io/projected/6aa59428-ee60-4434-9245-3ff8fa55200e-kube-api-access-fb55z\") pod \"6aa59428-ee60-4434-9245-3ff8fa55200e\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869347 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-plugins\") pod \"96acf88b-3fe8-46ac-8393-37f720632dd6\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869371 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-config\") pod \"6aa59428-ee60-4434-9245-3ff8fa55200e\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869390 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/96acf88b-3fe8-46ac-8393-37f720632dd6-pod-info\") pod \"96acf88b-3fe8-46ac-8393-37f720632dd6\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869410 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-erlang-cookie\") pod \"96acf88b-3fe8-46ac-8393-37f720632dd6\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869426 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-rundir\") pod \"6aa59428-ee60-4434-9245-3ff8fa55200e\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869448 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-northd-tls-certs\") pod \"6aa59428-ee60-4434-9245-3ff8fa55200e\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869480 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data\") pod \"96acf88b-3fe8-46ac-8393-37f720632dd6\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869513 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-tls\") pod \"96acf88b-3fe8-46ac-8393-37f720632dd6\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869530 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-server-conf\") pod \"96acf88b-3fe8-46ac-8393-37f720632dd6\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869565 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"96acf88b-3fe8-46ac-8393-37f720632dd6\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869594 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-combined-ca-bundle\") pod \"6aa59428-ee60-4434-9245-3ff8fa55200e\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869638 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-metrics-certs-tls-certs\") pod \"6aa59428-ee60-4434-9245-3ff8fa55200e\" (UID: \"6aa59428-ee60-4434-9245-3ff8fa55200e\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869657 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/96acf88b-3fe8-46ac-8393-37f720632dd6-erlang-cookie-secret\") pod \"96acf88b-3fe8-46ac-8393-37f720632dd6\" (UID: \"96acf88b-3fe8-46ac-8393-37f720632dd6\") " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869476 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "96acf88b-3fe8-46ac-8393-37f720632dd6" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869778 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-scripts" (OuterVolumeSpecName: "scripts") pod "6aa59428-ee60-4434-9245-3ff8fa55200e" (UID: "6aa59428-ee60-4434-9245-3ff8fa55200e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869912 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/366b7548-2fca-4623-a45c-c1e0367ce93a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869926 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869935 4799 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.869943 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.870312 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "96acf88b-3fe8-46ac-8393-37f720632dd6" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.870831 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "96acf88b-3fe8-46ac-8393-37f720632dd6" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.870903 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-config" (OuterVolumeSpecName: "config") pod "6aa59428-ee60-4434-9245-3ff8fa55200e" (UID: "6aa59428-ee60-4434-9245-3ff8fa55200e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.871011 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "6aa59428-ee60-4434-9245-3ff8fa55200e" (UID: "6aa59428-ee60-4434-9245-3ff8fa55200e"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.890608 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "96acf88b-3fe8-46ac-8393-37f720632dd6" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.892659 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-kube-api-access-wh8mr" (OuterVolumeSpecName: "kube-api-access-wh8mr") pod "96acf88b-3fe8-46ac-8393-37f720632dd6" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6"). InnerVolumeSpecName "kube-api-access-wh8mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.896456 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6aa59428-ee60-4434-9245-3ff8fa55200e" (UID: "6aa59428-ee60-4434-9245-3ff8fa55200e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.896987 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data" (OuterVolumeSpecName: "config-data") pod "96acf88b-3fe8-46ac-8393-37f720632dd6" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.897782 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aa59428-ee60-4434-9245-3ff8fa55200e-kube-api-access-fb55z" (OuterVolumeSpecName: "kube-api-access-fb55z") pod "6aa59428-ee60-4434-9245-3ff8fa55200e" (UID: "6aa59428-ee60-4434-9245-3ff8fa55200e"). InnerVolumeSpecName "kube-api-access-fb55z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.897822 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96acf88b-3fe8-46ac-8393-37f720632dd6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "96acf88b-3fe8-46ac-8393-37f720632dd6" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.898168 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/96acf88b-3fe8-46ac-8393-37f720632dd6-pod-info" (OuterVolumeSpecName: "pod-info") pod "96acf88b-3fe8-46ac-8393-37f720632dd6" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.898335 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "96acf88b-3fe8-46ac-8393-37f720632dd6" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.938178 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-server-conf" (OuterVolumeSpecName: "server-conf") pod "96acf88b-3fe8-46ac-8393-37f720632dd6" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.956116 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "6aa59428-ee60-4434-9245-3ff8fa55200e" (UID: "6aa59428-ee60-4434-9245-3ff8fa55200e"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.957096 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "6aa59428-ee60-4434-9245-3ff8fa55200e" (UID: "6aa59428-ee60-4434-9245-3ff8fa55200e"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971656 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh8mr\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-kube-api-access-wh8mr\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971714 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb55z\" (UniqueName: \"kubernetes.io/projected/6aa59428-ee60-4434-9245-3ff8fa55200e-kube-api-access-fb55z\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971729 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971742 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aa59428-ee60-4434-9245-3ff8fa55200e-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971751 4799 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/96acf88b-3fe8-46ac-8393-37f720632dd6-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971760 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971770 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971779 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971788 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971803 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971812 4799 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/96acf88b-3fe8-46ac-8393-37f720632dd6-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971863 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971875 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971890 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6aa59428-ee60-4434-9245-3ff8fa55200e-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.971902 4799 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/96acf88b-3fe8-46ac-8393-37f720632dd6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.989582 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "96acf88b-3fe8-46ac-8393-37f720632dd6" (UID: "96acf88b-3fe8-46ac-8393-37f720632dd6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:18 crc kubenswrapper[4799]: I1124 07:09:18.991754 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.073447 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.073481 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/96acf88b-3fe8-46ac-8393-37f720632dd6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.324106 4799 generic.go:334] "Generic (PLEG): container finished" podID="7603a96c-19e2-48ce-9c5f-69eeacc8cd85" containerID="29a4295f9247968819fb9ab51fb68ba67eeb2fb13b12fb980c93984bfc24c9e9" exitCode=0 Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.324193 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b94787bcf-mts4z" event={"ID":"7603a96c-19e2-48ce-9c5f-69eeacc8cd85","Type":"ContainerDied","Data":"29a4295f9247968819fb9ab51fb68ba67eeb2fb13b12fb980c93984bfc24c9e9"} Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.375183 4799 generic.go:334] "Generic (PLEG): container finished" podID="96acf88b-3fe8-46ac-8393-37f720632dd6" containerID="2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca" exitCode=0 Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.375358 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.375451 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"96acf88b-3fe8-46ac-8393-37f720632dd6","Type":"ContainerDied","Data":"2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca"} Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.375504 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"96acf88b-3fe8-46ac-8393-37f720632dd6","Type":"ContainerDied","Data":"be3cdda108578f8a4d7197ec7c3ca0efe2b97926c61c400bde868bb644f21f36"} Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.375525 4799 scope.go:117] "RemoveContainer" containerID="2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.428026 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6aa59428-ee60-4434-9245-3ff8fa55200e/ovn-northd/0.log" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.428078 4799 generic.go:334] "Generic (PLEG): container finished" podID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerID="2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1" exitCode=139 Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.428136 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6aa59428-ee60-4434-9245-3ff8fa55200e","Type":"ContainerDied","Data":"2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1"} Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.428164 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6aa59428-ee60-4434-9245-3ff8fa55200e","Type":"ContainerDied","Data":"975d6d84ae0a753e4c63ead344e1f49709456a53d737fca911099c1d0005019c"} Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.428261 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.434944 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"366b7548-2fca-4623-a45c-c1e0367ce93a","Type":"ContainerDied","Data":"602ba2b5867b8495e6b82dde9d4cfe4dedbf52572222ca4badcb957c182c94c2"} Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.435026 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.459950 4799 scope.go:117] "RemoveContainer" containerID="0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.513039 4799 scope.go:117] "RemoveContainer" containerID="2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.513269 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.516119 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca\": container with ID starting with 2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca not found: ID does not exist" containerID="2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.516855 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca"} err="failed to get container status \"2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca\": rpc error: code = NotFound desc = could not find container \"2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca\": container with ID starting with 2b0a6d939f82f6c742697904b5a73ce6f8e01e2237ee994e142e2a0a69b2c1ca not found: ID does not exist" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.516894 4799 scope.go:117] "RemoveContainer" containerID="0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29" Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.524600 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29\": container with ID starting with 0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29 not found: ID does not exist" containerID="0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.525342 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29"} err="failed to get container status \"0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29\": rpc error: code = NotFound desc = could not find container \"0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29\": container with ID starting with 0fa128cec1699ce614450603421f385d1ffcc3dd32849149542ef5ebff707c29 not found: ID does not exist" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.525779 4799 scope.go:117] "RemoveContainer" containerID="8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.544900 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.565958 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.573942 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.577258 4799 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 24 07:09:19 crc kubenswrapper[4799]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-11-24T07:09:12Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 24 07:09:19 crc kubenswrapper[4799]: /etc/init.d/functions: line 589: 379 Alarm clock "$@" Nov 24 07:09:19 crc kubenswrapper[4799]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-2dz2n" message=< Nov 24 07:09:19 crc kubenswrapper[4799]: Exiting ovn-controller (1) [FAILED] Nov 24 07:09:19 crc kubenswrapper[4799]: Killing ovn-controller (1) [ OK ] Nov 24 07:09:19 crc kubenswrapper[4799]: Killing ovn-controller (1) with SIGKILL [ OK ] Nov 24 07:09:19 crc kubenswrapper[4799]: 2025-11-24T07:09:12Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 24 07:09:19 crc kubenswrapper[4799]: /etc/init.d/functions: line 589: 379 Alarm clock "$@" Nov 24 07:09:19 crc kubenswrapper[4799]: > Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.577688 4799 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 24 07:09:19 crc kubenswrapper[4799]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-11-24T07:09:12Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 24 07:09:19 crc kubenswrapper[4799]: /etc/init.d/functions: line 589: 379 Alarm clock "$@" Nov 24 07:09:19 crc kubenswrapper[4799]: > pod="openstack/ovn-controller-2dz2n" podUID="dd32bea6-7ad9-4182-bc59-5644efe24ce1" containerName="ovn-controller" containerID="cri-o://5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.578054 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-2dz2n" podUID="dd32bea6-7ad9-4182-bc59-5644efe24ce1" containerName="ovn-controller" containerID="cri-o://5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983" gracePeriod=22 Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.581114 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.593254 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.601060 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.601143 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts podName:d13b66c3-1e31-418e-a0ff-0d1bf7a5980a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:23.601120901 +0000 UTC m=+1309.257103375 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts") pod "barbican44f2-account-delete-2hm9q" (UID: "d13b66c3-1e31-418e-a0ff-0d1bf7a5980a") : configmap "openstack-scripts" not found Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.601400 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.601482 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts podName:87439e47-6c84-4467-a49b-03ed6e86a9eb nodeName:}" failed. No retries permitted until 2025-11-24 07:09:23.601462531 +0000 UTC m=+1309.257444995 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts") pod "novacell000c8-account-delete-n5mhs" (UID: "87439e47-6c84-4467-a49b-03ed6e86a9eb") : configmap "openstack-scripts" not found Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.649189 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="168550d3-6597-41ff-adf4-d20f190c7a02" path="/var/lib/kubelet/pods/168550d3-6597-41ff-adf4-d20f190c7a02/volumes" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.652825 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="366b7548-2fca-4623-a45c-c1e0367ce93a" path="/var/lib/kubelet/pods/366b7548-2fca-4623-a45c-c1e0367ce93a/volumes" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.653527 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63757cd4-2ed0-4423-b466-4dfac76e3ac4" path="/var/lib/kubelet/pods/63757cd4-2ed0-4423-b466-4dfac76e3ac4/volumes" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.659078 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aa59428-ee60-4434-9245-3ff8fa55200e" path="/var/lib/kubelet/pods/6aa59428-ee60-4434-9245-3ff8fa55200e/volumes" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.660275 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" path="/var/lib/kubelet/pods/6fe03a76-2eb9-4728-9396-9d2dc1e743af/volumes" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.661035 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96acf88b-3fe8-46ac-8393-37f720632dd6" path="/var/lib/kubelet/pods/96acf88b-3fe8-46ac-8393-37f720632dd6/volumes" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.663022 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="997ffb9e-cc6a-4fca-91d9-eb17d711b135" path="/var/lib/kubelet/pods/997ffb9e-cc6a-4fca-91d9-eb17d711b135/volumes" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.663610 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb94fabc-dfc3-40cc-9791-0272924692da" path="/var/lib/kubelet/pods/cb94fabc-dfc3-40cc-9791-0272924692da/volumes" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.665269 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebfe77df-bbcd-412f-9703-78e23d93f410" path="/var/lib/kubelet/pods/ebfe77df-bbcd-412f-9703-78e23d93f410/volumes" Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.678093 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6aa59428_ee60_4434_9245_3ff8fa55200e.slice/crio-975d6d84ae0a753e4c63ead344e1f49709456a53d737fca911099c1d0005019c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod366b7548_2fca_4623_a45c_c1e0367ce93a.slice/crio-602ba2b5867b8495e6b82dde9d4cfe4dedbf52572222ca4badcb957c182c94c2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96acf88b_3fe8_46ac_8393_37f720632dd6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6aa59428_ee60_4434_9245_3ff8fa55200e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod366b7548_2fca_4623_a45c_c1e0367ce93a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd32bea6_7ad9_4182_bc59_5644efe24ce1.slice/crio-5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983.scope\": RecentStats: unable to find data in memory cache]" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.701609 4799 scope.go:117] "RemoveContainer" containerID="2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1" Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.702098 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.702186 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts podName:b6f22313-6d9a-4b35-b892-695903d354b1 nodeName:}" failed. No retries permitted until 2025-11-24 07:09:23.702170529 +0000 UTC m=+1309.358153003 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts") pod "placement2704-account-delete-ck9t2" (UID: "b6f22313-6d9a-4b35-b892-695903d354b1") : configmap "openstack-scripts" not found Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.706481 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.750068 4799 scope.go:117] "RemoveContainer" containerID="8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80" Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.754029 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80\": container with ID starting with 8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80 not found: ID does not exist" containerID="8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.754093 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80"} err="failed to get container status \"8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80\": rpc error: code = NotFound desc = could not find container \"8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80\": container with ID starting with 8cf025878bc3fb8e4437787385f0db356cc0b01c4e95f3224426f578bb82ca80 not found: ID does not exist" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.754128 4799 scope.go:117] "RemoveContainer" containerID="2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1" Nov 24 07:09:19 crc kubenswrapper[4799]: E1124 07:09:19.756226 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1\": container with ID starting with 2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1 not found: ID does not exist" containerID="2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.756276 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1"} err="failed to get container status \"2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1\": rpc error: code = NotFound desc = could not find container \"2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1\": container with ID starting with 2ea81447c7a07862cdd22958c70a2c86084d0a837c13cd65af82450ce4ef48e1 not found: ID does not exist" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.756305 4799 scope.go:117] "RemoveContainer" containerID="54abd42b7590aea7f2db6c595fca4f645386857fcbdb38711dfb864deb083510" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.799533 4799 scope.go:117] "RemoveContainer" containerID="c6e22b9dd219757023e664a05027fc4af7ee03f3c7e62fe2a381ed00aef63570" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.807674 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-public-tls-certs\") pod \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.807725 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-config-data\") pod \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.807745 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-fernet-keys\") pod \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.807767 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-internal-tls-certs\") pod \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.807869 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-credential-keys\") pod \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.807890 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-combined-ca-bundle\") pod \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.807916 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr8lx\" (UniqueName: \"kubernetes.io/projected/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-kube-api-access-lr8lx\") pod \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.807945 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-scripts\") pod \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\" (UID: \"7603a96c-19e2-48ce-9c5f-69eeacc8cd85\") " Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.812087 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7603a96c-19e2-48ce-9c5f-69eeacc8cd85" (UID: "7603a96c-19e2-48ce-9c5f-69eeacc8cd85"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.817043 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-kube-api-access-lr8lx" (OuterVolumeSpecName: "kube-api-access-lr8lx") pod "7603a96c-19e2-48ce-9c5f-69eeacc8cd85" (UID: "7603a96c-19e2-48ce-9c5f-69eeacc8cd85"). InnerVolumeSpecName "kube-api-access-lr8lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.818715 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-scripts" (OuterVolumeSpecName: "scripts") pod "7603a96c-19e2-48ce-9c5f-69eeacc8cd85" (UID: "7603a96c-19e2-48ce-9c5f-69eeacc8cd85"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.838955 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7603a96c-19e2-48ce-9c5f-69eeacc8cd85" (UID: "7603a96c-19e2-48ce-9c5f-69eeacc8cd85"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.872332 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-v4ct4"] Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.895707 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-v4ct4"] Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.910223 4799 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.910253 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr8lx\" (UniqueName: \"kubernetes.io/projected/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-kube-api-access-lr8lx\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.910265 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.910274 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.917811 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderbbb1-account-delete-tjfp2"] Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.921989 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7603a96c-19e2-48ce-9c5f-69eeacc8cd85" (UID: "7603a96c-19e2-48ce-9c5f-69eeacc8cd85"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.922082 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7603a96c-19e2-48ce-9c5f-69eeacc8cd85" (UID: "7603a96c-19e2-48ce-9c5f-69eeacc8cd85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.923691 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7603a96c-19e2-48ce-9c5f-69eeacc8cd85" (UID: "7603a96c-19e2-48ce-9c5f-69eeacc8cd85"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.927876 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-config-data" (OuterVolumeSpecName: "config-data") pod "7603a96c-19e2-48ce-9c5f-69eeacc8cd85" (UID: "7603a96c-19e2-48ce-9c5f-69eeacc8cd85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.937921 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-bbb1-account-create-87l6k"] Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.946438 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-bbb1-account-create-87l6k"] Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.953539 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-2dz2n_dd32bea6-7ad9-4182-bc59-5644efe24ce1/ovn-controller/0.log" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.953722 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n" Nov 24 07:09:19 crc kubenswrapper[4799]: I1124 07:09:19.955775 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinderbbb1-account-delete-tjfp2"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.001168 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-b5qgf"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.012905 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.012929 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.012937 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.012945 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7603a96c-19e2-48ce-9c5f-69eeacc8cd85-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.027776 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-b5qgf"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.036361 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-00d0-account-create-xqldv"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.045770 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance00d0-account-delete-kbtpl"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.052665 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-00d0-account-create-xqldv"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.057085 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance00d0-account-delete-kbtpl"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.113555 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd32bea6-7ad9-4182-bc59-5644efe24ce1-scripts\") pod \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.113594 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-combined-ca-bundle\") pod \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.113639 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run-ovn\") pod \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.113668 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jrkq\" (UniqueName: \"kubernetes.io/projected/dd32bea6-7ad9-4182-bc59-5644efe24ce1-kube-api-access-6jrkq\") pod \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.113724 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run\") pod \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.113749 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "dd32bea6-7ad9-4182-bc59-5644efe24ce1" (UID: "dd32bea6-7ad9-4182-bc59-5644efe24ce1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.113817 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-ovn-controller-tls-certs\") pod \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.113837 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-log-ovn\") pod \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\" (UID: \"dd32bea6-7ad9-4182-bc59-5644efe24ce1\") " Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.113923 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "dd32bea6-7ad9-4182-bc59-5644efe24ce1" (UID: "dd32bea6-7ad9-4182-bc59-5644efe24ce1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.113981 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run" (OuterVolumeSpecName: "var-run") pod "dd32bea6-7ad9-4182-bc59-5644efe24ce1" (UID: "dd32bea6-7ad9-4182-bc59-5644efe24ce1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.114293 4799 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.114308 4799 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.114316 4799 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/dd32bea6-7ad9-4182-bc59-5644efe24ce1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.115294 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd32bea6-7ad9-4182-bc59-5644efe24ce1-scripts" (OuterVolumeSpecName: "scripts") pod "dd32bea6-7ad9-4182-bc59-5644efe24ce1" (UID: "dd32bea6-7ad9-4182-bc59-5644efe24ce1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.121037 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd32bea6-7ad9-4182-bc59-5644efe24ce1-kube-api-access-6jrkq" (OuterVolumeSpecName: "kube-api-access-6jrkq") pod "dd32bea6-7ad9-4182-bc59-5644efe24ce1" (UID: "dd32bea6-7ad9-4182-bc59-5644efe24ce1"). InnerVolumeSpecName "kube-api-access-6jrkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.149250 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd32bea6-7ad9-4182-bc59-5644efe24ce1" (UID: "dd32bea6-7ad9-4182-bc59-5644efe24ce1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.215891 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd32bea6-7ad9-4182-bc59-5644efe24ce1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.215930 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.215970 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jrkq\" (UniqueName: \"kubernetes.io/projected/dd32bea6-7ad9-4182-bc59-5644efe24ce1-kube-api-access-6jrkq\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: E1124 07:09:20.217784 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:20 crc kubenswrapper[4799]: E1124 07:09:20.218757 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:20 crc kubenswrapper[4799]: E1124 07:09:20.220838 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:20 crc kubenswrapper[4799]: E1124 07:09:20.220895 4799 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.223429 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "dd32bea6-7ad9-4182-bc59-5644efe24ce1" (UID: "dd32bea6-7ad9-4182-bc59-5644efe24ce1"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.233246 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-hslpd"] Nov 24 07:09:20 crc kubenswrapper[4799]: E1124 07:09:20.240521 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:20 crc kubenswrapper[4799]: E1124 07:09:20.242698 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:20 crc kubenswrapper[4799]: E1124 07:09:20.246251 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:20 crc kubenswrapper[4799]: E1124 07:09:20.246303 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovs-vswitchd" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.263650 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-hslpd"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.276787 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron2d2f-account-delete-5c2sd"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.283798 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2d2f-account-create-mgts2"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.294984 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2d2f-account-create-mgts2"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.296803 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron2d2f-account-delete-5c2sd"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.301474 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-v9jnb"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.306597 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-v9jnb"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.316993 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican44f2-account-delete-2hm9q"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.317266 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican44f2-account-delete-2hm9q" podUID="d13b66c3-1e31-418e-a0ff-0d1bf7a5980a" containerName="mariadb-account-delete" containerID="cri-o://16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090" gracePeriod=30 Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.318453 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd32bea6-7ad9-4182-bc59-5644efe24ce1-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.321965 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-44f2-account-create-ds5c4"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.327107 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-44f2-account-create-ds5c4"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.401511 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.401583 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.406029 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.455774 4799 generic.go:334] "Generic (PLEG): container finished" podID="66ae5dfe-6c96-4d2d-99cf-b792b139d1a1" containerID="8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05" exitCode=0 Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.455834 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.455919 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1","Type":"ContainerDied","Data":"8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05"} Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.455953 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1","Type":"ContainerDied","Data":"1d129c757ca1658503bc21c474818337e63f0f84ff4d13fb65c9e13277d1a15c"} Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.455971 4799 scope.go:117] "RemoveContainer" containerID="8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.459899 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-2dz2n_dd32bea6-7ad9-4182-bc59-5644efe24ce1/ovn-controller/0.log" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.459946 4799 generic.go:334] "Generic (PLEG): container finished" podID="dd32bea6-7ad9-4182-bc59-5644efe24ce1" containerID="5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983" exitCode=137 Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.460011 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n" event={"ID":"dd32bea6-7ad9-4182-bc59-5644efe24ce1","Type":"ContainerDied","Data":"5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983"} Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.460035 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2dz2n" event={"ID":"dd32bea6-7ad9-4182-bc59-5644efe24ce1","Type":"ContainerDied","Data":"ca0366e7a6e2436619aa02f7351d51fd3452b3bdb2ad73621e1c3046ac9d75e3"} Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.460047 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2dz2n" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.467160 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b94787bcf-mts4z" event={"ID":"7603a96c-19e2-48ce-9c5f-69eeacc8cd85","Type":"ContainerDied","Data":"60197a8ca0e4d3646b4e4bd496c61312a49d9782b200ab25fbbb3989c5ad162d"} Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.467262 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b94787bcf-mts4z" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.488905 4799 scope.go:117] "RemoveContainer" containerID="8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05" Nov 24 07:09:20 crc kubenswrapper[4799]: E1124 07:09:20.489516 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05\": container with ID starting with 8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05 not found: ID does not exist" containerID="8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.489668 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05"} err="failed to get container status \"8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05\": rpc error: code = NotFound desc = could not find container \"8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05\": container with ID starting with 8ee3bdf3a575c9b4a13ae1c10603336a2dbd7c2330a846e2c13bfb690b2a1d05 not found: ID does not exist" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.489767 4799 scope.go:117] "RemoveContainer" containerID="5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.524039 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7nqh\" (UniqueName: \"kubernetes.io/projected/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-kube-api-access-s7nqh\") pod \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.524277 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-config-data\") pod \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.524340 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-combined-ca-bundle\") pod \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\" (UID: \"66ae5dfe-6c96-4d2d-99cf-b792b139d1a1\") " Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.531394 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-2dz2n"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.531931 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-kube-api-access-s7nqh" (OuterVolumeSpecName: "kube-api-access-s7nqh") pod "66ae5dfe-6c96-4d2d-99cf-b792b139d1a1" (UID: "66ae5dfe-6c96-4d2d-99cf-b792b139d1a1"). InnerVolumeSpecName "kube-api-access-s7nqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.539817 4799 scope.go:117] "RemoveContainer" containerID="5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983" Nov 24 07:09:20 crc kubenswrapper[4799]: E1124 07:09:20.541081 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983\": container with ID starting with 5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983 not found: ID does not exist" containerID="5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.541177 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983"} err="failed to get container status \"5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983\": rpc error: code = NotFound desc = could not find container \"5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983\": container with ID starting with 5d914939d250aaf9efc0b4fd78bd9ad2e4f12fa15661230165d4ec39f9be1983 not found: ID does not exist" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.541244 4799 scope.go:117] "RemoveContainer" containerID="29a4295f9247968819fb9ab51fb68ba67eeb2fb13b12fb980c93984bfc24c9e9" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.541453 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-2dz2n"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.561790 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66ae5dfe-6c96-4d2d-99cf-b792b139d1a1" (UID: "66ae5dfe-6c96-4d2d-99cf-b792b139d1a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.577553 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-config-data" (OuterVolumeSpecName: "config-data") pod "66ae5dfe-6c96-4d2d-99cf-b792b139d1a1" (UID: "66ae5dfe-6c96-4d2d-99cf-b792b139d1a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.580587 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7b94787bcf-mts4z"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.587668 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7b94787bcf-mts4z"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.627220 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.627528 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.627606 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7nqh\" (UniqueName: \"kubernetes.io/projected/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1-kube-api-access-s7nqh\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.702224 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="3cd44457-1418-4788-a029-82c54c917997" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.194:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.791067 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 07:09:20 crc kubenswrapper[4799]: I1124 07:09:20.798203 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.641220 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1331194e-76cf-4c49-9d14-294d7b3897bd" path="/var/lib/kubelet/pods/1331194e-76cf-4c49-9d14-294d7b3897bd/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.642383 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="207abd4d-7482-4d1e-93b8-abab45c55888" path="/var/lib/kubelet/pods/207abd4d-7482-4d1e-93b8-abab45c55888/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.643731 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="644e0295-5410-47b5-8cb0-ffc0f89d8b13" path="/var/lib/kubelet/pods/644e0295-5410-47b5-8cb0-ffc0f89d8b13/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.644932 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66ae5dfe-6c96-4d2d-99cf-b792b139d1a1" path="/var/lib/kubelet/pods/66ae5dfe-6c96-4d2d-99cf-b792b139d1a1/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.646880 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7603a96c-19e2-48ce-9c5f-69eeacc8cd85" path="/var/lib/kubelet/pods/7603a96c-19e2-48ce-9c5f-69eeacc8cd85/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.648072 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99c90963-48b3-45aa-85cc-f6d51b5702be" path="/var/lib/kubelet/pods/99c90963-48b3-45aa-85cc-f6d51b5702be/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.649072 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3" path="/var/lib/kubelet/pods/a8e68fa0-0bd0-4c7b-93b3-7b8cb2f7f7b3/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.651061 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9e7a595-81bd-4b15-8682-fbdf8ab08e5e" path="/var/lib/kubelet/pods/a9e7a595-81bd-4b15-8682-fbdf8ab08e5e/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.652090 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad75f933-ab36-4c9d-857d-c2cc11702dd3" path="/var/lib/kubelet/pods/ad75f933-ab36-4c9d-857d-c2cc11702dd3/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.653050 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f86cb1-3f0e-4da8-82dd-e00522d4ceee" path="/var/lib/kubelet/pods/d3f86cb1-3f0e-4da8-82dd-e00522d4ceee/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.655064 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9064d19-6806-4e78-beb4-cda924fe3010" path="/var/lib/kubelet/pods/d9064d19-6806-4e78-beb4-cda924fe3010/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.656222 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd32bea6-7ad9-4182-bc59-5644efe24ce1" path="/var/lib/kubelet/pods/dd32bea6-7ad9-4182-bc59-5644efe24ce1/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.657226 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e00ef351-7e26-4b68-8c9b-a37510eff377" path="/var/lib/kubelet/pods/e00ef351-7e26-4b68-8c9b-a37510eff377/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.658269 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0c34246-d1bb-401c-94e5-084d6064ce5b" path="/var/lib/kubelet/pods/e0c34246-d1bb-401c-94e5-084d6064ce5b/volumes" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.700139 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-644b86779b-b4psv" podUID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.161:9311/healthcheck\": context deadline exceeded" Nov 24 07:09:21 crc kubenswrapper[4799]: I1124 07:09:21.700133 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-644b86779b-b4psv" podUID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.161:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 07:09:22 crc kubenswrapper[4799]: I1124 07:09:22.545964 4799 generic.go:334] "Generic (PLEG): container finished" podID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerID="247ac76988e78afd0fb67b14cc28cf0d059ad2c2f3abbf816f4a4d4fb2204cb3" exitCode=0 Nov 24 07:09:22 crc kubenswrapper[4799]: I1124 07:09:22.546422 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f97d9a56-cb78-4da9-b832-0028b6e45895","Type":"ContainerDied","Data":"247ac76988e78afd0fb67b14cc28cf0d059ad2c2f3abbf816f4a4d4fb2204cb3"} Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.755991 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.884294 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-ceilometer-tls-certs\") pod \"f97d9a56-cb78-4da9-b832-0028b6e45895\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.884364 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-sg-core-conf-yaml\") pod \"f97d9a56-cb78-4da9-b832-0028b6e45895\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.884407 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-log-httpd\") pod \"f97d9a56-cb78-4da9-b832-0028b6e45895\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.884430 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-scripts\") pod \"f97d9a56-cb78-4da9-b832-0028b6e45895\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.884488 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-combined-ca-bundle\") pod \"f97d9a56-cb78-4da9-b832-0028b6e45895\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.884531 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-config-data\") pod \"f97d9a56-cb78-4da9-b832-0028b6e45895\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.884575 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4xlw\" (UniqueName: \"kubernetes.io/projected/f97d9a56-cb78-4da9-b832-0028b6e45895-kube-api-access-q4xlw\") pod \"f97d9a56-cb78-4da9-b832-0028b6e45895\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.884645 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-run-httpd\") pod \"f97d9a56-cb78-4da9-b832-0028b6e45895\" (UID: \"f97d9a56-cb78-4da9-b832-0028b6e45895\") " Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.885414 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f97d9a56-cb78-4da9-b832-0028b6e45895" (UID: "f97d9a56-cb78-4da9-b832-0028b6e45895"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.885452 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f97d9a56-cb78-4da9-b832-0028b6e45895" (UID: "f97d9a56-cb78-4da9-b832-0028b6e45895"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.891220 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-scripts" (OuterVolumeSpecName: "scripts") pod "f97d9a56-cb78-4da9-b832-0028b6e45895" (UID: "f97d9a56-cb78-4da9-b832-0028b6e45895"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.891890 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f97d9a56-cb78-4da9-b832-0028b6e45895-kube-api-access-q4xlw" (OuterVolumeSpecName: "kube-api-access-q4xlw") pod "f97d9a56-cb78-4da9-b832-0028b6e45895" (UID: "f97d9a56-cb78-4da9-b832-0028b6e45895"). InnerVolumeSpecName "kube-api-access-q4xlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.939246 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f97d9a56-cb78-4da9-b832-0028b6e45895" (UID: "f97d9a56-cb78-4da9-b832-0028b6e45895"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.942139 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f97d9a56-cb78-4da9-b832-0028b6e45895" (UID: "f97d9a56-cb78-4da9-b832-0028b6e45895"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.978098 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f97d9a56-cb78-4da9-b832-0028b6e45895" (UID: "f97d9a56-cb78-4da9-b832-0028b6e45895"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.986373 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.986424 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.986434 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.986446 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.986457 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.986467 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4xlw\" (UniqueName: \"kubernetes.io/projected/f97d9a56-cb78-4da9-b832-0028b6e45895-kube-api-access-q4xlw\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.986478 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f97d9a56-cb78-4da9-b832-0028b6e45895-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:22.992313 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-config-data" (OuterVolumeSpecName: "config-data") pod "f97d9a56-cb78-4da9-b832-0028b6e45895" (UID: "f97d9a56-cb78-4da9-b832-0028b6e45895"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:23.088634 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f97d9a56-cb78-4da9-b832-0028b6e45895-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:23.564400 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f97d9a56-cb78-4da9-b832-0028b6e45895","Type":"ContainerDied","Data":"1210aa6be7227456168aeae1f4893423c3e858db424c8c026e2f2703c17d5d42"} Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:23.565100 4799 scope.go:117] "RemoveContainer" containerID="dc5fdc4d831816407f510ddb2a4fc3bcb5bbc530983dfb627e98b50af968d2bd" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:23.564473 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 07:09:23 crc kubenswrapper[4799]: E1124 07:09:23.602697 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:23 crc kubenswrapper[4799]: E1124 07:09:23.602964 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts podName:87439e47-6c84-4467-a49b-03ed6e86a9eb nodeName:}" failed. No retries permitted until 2025-11-24 07:09:31.602835859 +0000 UTC m=+1317.258818363 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts") pod "novacell000c8-account-delete-n5mhs" (UID: "87439e47-6c84-4467-a49b-03ed6e86a9eb") : configmap "openstack-scripts" not found Nov 24 07:09:23 crc kubenswrapper[4799]: E1124 07:09:23.603302 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:23 crc kubenswrapper[4799]: E1124 07:09:23.603436 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts podName:d13b66c3-1e31-418e-a0ff-0d1bf7a5980a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:31.603402215 +0000 UTC m=+1317.259384869 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts") pod "barbican44f2-account-delete-2hm9q" (UID: "d13b66c3-1e31-418e-a0ff-0d1bf7a5980a") : configmap "openstack-scripts" not found Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:23.607233 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:23.609689 4799 scope.go:117] "RemoveContainer" containerID="6b464ae1867d289634184c0a2a89b04d292707252e13139c622836b8234cc5e5" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:23.622417 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:23.636307 4799 scope.go:117] "RemoveContainer" containerID="247ac76988e78afd0fb67b14cc28cf0d059ad2c2f3abbf816f4a4d4fb2204cb3" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:23.641080 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" path="/var/lib/kubelet/pods/f97d9a56-cb78-4da9-b832-0028b6e45895/volumes" Nov 24 07:09:23 crc kubenswrapper[4799]: I1124 07:09:23.669438 4799 scope.go:117] "RemoveContainer" containerID="4b1017dff22faa9e7cf62359f3de7ceea284aa7bf591e9d872e15ff649d22d96" Nov 24 07:09:23 crc kubenswrapper[4799]: E1124 07:09:23.706318 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:23 crc kubenswrapper[4799]: E1124 07:09:23.706931 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts podName:b6f22313-6d9a-4b35-b892-695903d354b1 nodeName:}" failed. No retries permitted until 2025-11-24 07:09:31.706901622 +0000 UTC m=+1317.362884106 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts") pod "placement2704-account-delete-ck9t2" (UID: "b6f22313-6d9a-4b35-b892-695903d354b1") : configmap "openstack-scripts" not found Nov 24 07:09:25 crc kubenswrapper[4799]: E1124 07:09:25.216977 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:25 crc kubenswrapper[4799]: E1124 07:09:25.218063 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:25 crc kubenswrapper[4799]: E1124 07:09:25.218255 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:25 crc kubenswrapper[4799]: E1124 07:09:25.219230 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:25 crc kubenswrapper[4799]: E1124 07:09:25.219298 4799 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server" Nov 24 07:09:25 crc kubenswrapper[4799]: E1124 07:09:25.219685 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:25 crc kubenswrapper[4799]: E1124 07:09:25.221615 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:25 crc kubenswrapper[4799]: E1124 07:09:25.221683 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovs-vswitchd" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.614343 4799 generic.go:334] "Generic (PLEG): container finished" podID="0a50ddac-d55d-4e99-8dea-8519b8d732d6" containerID="02ffa0bb3d383e7c692b859a1e3376a9f3f7187ae40a809812908dbb07a5603c" exitCode=0 Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.614567 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77fc689cdf-xj7m6" event={"ID":"0a50ddac-d55d-4e99-8dea-8519b8d732d6","Type":"ContainerDied","Data":"02ffa0bb3d383e7c692b859a1e3376a9f3f7187ae40a809812908dbb07a5603c"} Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.614708 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77fc689cdf-xj7m6" event={"ID":"0a50ddac-d55d-4e99-8dea-8519b8d732d6","Type":"ContainerDied","Data":"bab175d3ca6e71c3354bcdb618c71fac89d3d23a05be126cef27869bd259e337"} Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.614725 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bab175d3ca6e71c3354bcdb618c71fac89d3d23a05be126cef27869bd259e337" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.640460 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.759896 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-public-tls-certs\") pod \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.759959 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-config\") pod \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.760007 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-httpd-config\") pod \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.760026 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjff2\" (UniqueName: \"kubernetes.io/projected/0a50ddac-d55d-4e99-8dea-8519b8d732d6-kube-api-access-qjff2\") pod \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.760102 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-ovndb-tls-certs\") pod \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.760139 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-internal-tls-certs\") pod \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.760206 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-combined-ca-bundle\") pod \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.768201 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a50ddac-d55d-4e99-8dea-8519b8d732d6-kube-api-access-qjff2" (OuterVolumeSpecName: "kube-api-access-qjff2") pod "0a50ddac-d55d-4e99-8dea-8519b8d732d6" (UID: "0a50ddac-d55d-4e99-8dea-8519b8d732d6"). InnerVolumeSpecName "kube-api-access-qjff2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.771935 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "0a50ddac-d55d-4e99-8dea-8519b8d732d6" (UID: "0a50ddac-d55d-4e99-8dea-8519b8d732d6"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.826568 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0a50ddac-d55d-4e99-8dea-8519b8d732d6" (UID: "0a50ddac-d55d-4e99-8dea-8519b8d732d6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.829013 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a50ddac-d55d-4e99-8dea-8519b8d732d6" (UID: "0a50ddac-d55d-4e99-8dea-8519b8d732d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.848736 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0a50ddac-d55d-4e99-8dea-8519b8d732d6" (UID: "0a50ddac-d55d-4e99-8dea-8519b8d732d6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.849120 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "0a50ddac-d55d-4e99-8dea-8519b8d732d6" (UID: "0a50ddac-d55d-4e99-8dea-8519b8d732d6"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.860996 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-config" (OuterVolumeSpecName: "config") pod "0a50ddac-d55d-4e99-8dea-8519b8d732d6" (UID: "0a50ddac-d55d-4e99-8dea-8519b8d732d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.861600 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-config\") pod \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\" (UID: \"0a50ddac-d55d-4e99-8dea-8519b8d732d6\") " Nov 24 07:09:26 crc kubenswrapper[4799]: W1124 07:09:26.861777 4799 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/0a50ddac-d55d-4e99-8dea-8519b8d732d6/volumes/kubernetes.io~secret/config Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.861803 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-config" (OuterVolumeSpecName: "config") pod "0a50ddac-d55d-4e99-8dea-8519b8d732d6" (UID: "0a50ddac-d55d-4e99-8dea-8519b8d732d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.863162 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.863194 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.863203 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.863214 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjff2\" (UniqueName: \"kubernetes.io/projected/0a50ddac-d55d-4e99-8dea-8519b8d732d6-kube-api-access-qjff2\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.863228 4799 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.863238 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:26 crc kubenswrapper[4799]: I1124 07:09:26.863247 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a50ddac-d55d-4e99-8dea-8519b8d732d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:27 crc kubenswrapper[4799]: I1124 07:09:27.631202 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77fc689cdf-xj7m6" Nov 24 07:09:27 crc kubenswrapper[4799]: I1124 07:09:27.697421 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-77fc689cdf-xj7m6"] Nov 24 07:09:27 crc kubenswrapper[4799]: I1124 07:09:27.704974 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-77fc689cdf-xj7m6"] Nov 24 07:09:29 crc kubenswrapper[4799]: I1124 07:09:29.654640 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a50ddac-d55d-4e99-8dea-8519b8d732d6" path="/var/lib/kubelet/pods/0a50ddac-d55d-4e99-8dea-8519b8d732d6/volumes" Nov 24 07:09:30 crc kubenswrapper[4799]: E1124 07:09:30.217205 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:30 crc kubenswrapper[4799]: E1124 07:09:30.218465 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:30 crc kubenswrapper[4799]: E1124 07:09:30.219393 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:30 crc kubenswrapper[4799]: E1124 07:09:30.219457 4799 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server" Nov 24 07:09:30 crc kubenswrapper[4799]: E1124 07:09:30.220326 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:30 crc kubenswrapper[4799]: E1124 07:09:30.223452 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:30 crc kubenswrapper[4799]: E1124 07:09:30.226264 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:30 crc kubenswrapper[4799]: E1124 07:09:30.226354 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovs-vswitchd" Nov 24 07:09:31 crc kubenswrapper[4799]: E1124 07:09:31.649563 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:31 crc kubenswrapper[4799]: E1124 07:09:31.649635 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:31 crc kubenswrapper[4799]: E1124 07:09:31.649655 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts podName:d13b66c3-1e31-418e-a0ff-0d1bf7a5980a nodeName:}" failed. No retries permitted until 2025-11-24 07:09:47.649632456 +0000 UTC m=+1333.305614920 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts") pod "barbican44f2-account-delete-2hm9q" (UID: "d13b66c3-1e31-418e-a0ff-0d1bf7a5980a") : configmap "openstack-scripts" not found Nov 24 07:09:31 crc kubenswrapper[4799]: E1124 07:09:31.649748 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts podName:87439e47-6c84-4467-a49b-03ed6e86a9eb nodeName:}" failed. No retries permitted until 2025-11-24 07:09:47.649722988 +0000 UTC m=+1333.305705602 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts") pod "novacell000c8-account-delete-n5mhs" (UID: "87439e47-6c84-4467-a49b-03ed6e86a9eb") : configmap "openstack-scripts" not found Nov 24 07:09:31 crc kubenswrapper[4799]: E1124 07:09:31.751017 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:31 crc kubenswrapper[4799]: E1124 07:09:31.751121 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts podName:b6f22313-6d9a-4b35-b892-695903d354b1 nodeName:}" failed. No retries permitted until 2025-11-24 07:09:47.751100815 +0000 UTC m=+1333.407083279 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts") pod "placement2704-account-delete-ck9t2" (UID: "b6f22313-6d9a-4b35-b892-695903d354b1") : configmap "openstack-scripts" not found Nov 24 07:09:35 crc kubenswrapper[4799]: E1124 07:09:35.216600 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:35 crc kubenswrapper[4799]: E1124 07:09:35.218318 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:35 crc kubenswrapper[4799]: E1124 07:09:35.218830 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:35 crc kubenswrapper[4799]: E1124 07:09:35.218999 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:35 crc kubenswrapper[4799]: E1124 07:09:35.219031 4799 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server" Nov 24 07:09:35 crc kubenswrapper[4799]: E1124 07:09:35.220982 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:35 crc kubenswrapper[4799]: E1124 07:09:35.222993 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:35 crc kubenswrapper[4799]: E1124 07:09:35.223040 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovs-vswitchd" Nov 24 07:09:40 crc kubenswrapper[4799]: E1124 07:09:40.217603 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:40 crc kubenswrapper[4799]: E1124 07:09:40.219276 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:40 crc kubenswrapper[4799]: E1124 07:09:40.223276 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:40 crc kubenswrapper[4799]: E1124 07:09:40.224145 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 07:09:40 crc kubenswrapper[4799]: E1124 07:09:40.224218 4799 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server" Nov 24 07:09:40 crc kubenswrapper[4799]: E1124 07:09:40.225074 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:40 crc kubenswrapper[4799]: E1124 07:09:40.227460 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 07:09:40 crc kubenswrapper[4799]: E1124 07:09:40.227491 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-kgrrs" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovs-vswitchd" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.804702 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kgrrs_a7742c81-45e0-4ef5-aceb-7a631bf9cc37/ovs-vswitchd/0.log" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.806201 4799 generic.go:334] "Generic (PLEG): container finished" podID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" exitCode=137 Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.806310 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kgrrs" event={"ID":"a7742c81-45e0-4ef5-aceb-7a631bf9cc37","Type":"ContainerDied","Data":"16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f"} Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.818968 4799 generic.go:334] "Generic (PLEG): container finished" podID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerID="57ac5adb075544c22ae8aa32fb5357e1f72bc9a58d99ba17c8bb148b66dade15" exitCode=137 Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.819017 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"57ac5adb075544c22ae8aa32fb5357e1f72bc9a58d99ba17c8bb148b66dade15"} Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.819157 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a30589ef-eff4-4113-8e6a-32f03be5c734","Type":"ContainerDied","Data":"c599b8c2be169b387750d307815794693bda4ae021f7c3c39ac452e297437fc2"} Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.819183 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c599b8c2be169b387750d307815794693bda4ae021f7c3c39ac452e297437fc2" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.826889 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.850746 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift\") pod \"a30589ef-eff4-4113-8e6a-32f03be5c734\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.851271 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-cache\") pod \"a30589ef-eff4-4113-8e6a-32f03be5c734\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.851354 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"a30589ef-eff4-4113-8e6a-32f03be5c734\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.851578 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckrjg\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-kube-api-access-ckrjg\") pod \"a30589ef-eff4-4113-8e6a-32f03be5c734\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.851787 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-lock\") pod \"a30589ef-eff4-4113-8e6a-32f03be5c734\" (UID: \"a30589ef-eff4-4113-8e6a-32f03be5c734\") " Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.852016 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-cache" (OuterVolumeSpecName: "cache") pod "a30589ef-eff4-4113-8e6a-32f03be5c734" (UID: "a30589ef-eff4-4113-8e6a-32f03be5c734"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.853810 4799 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-cache\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.854612 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-lock" (OuterVolumeSpecName: "lock") pod "a30589ef-eff4-4113-8e6a-32f03be5c734" (UID: "a30589ef-eff4-4113-8e6a-32f03be5c734"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.863200 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a30589ef-eff4-4113-8e6a-32f03be5c734" (UID: "a30589ef-eff4-4113-8e6a-32f03be5c734"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.863256 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-kube-api-access-ckrjg" (OuterVolumeSpecName: "kube-api-access-ckrjg") pod "a30589ef-eff4-4113-8e6a-32f03be5c734" (UID: "a30589ef-eff4-4113-8e6a-32f03be5c734"). InnerVolumeSpecName "kube-api-access-ckrjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.867576 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "swift") pod "a30589ef-eff4-4113-8e6a-32f03be5c734" (UID: "a30589ef-eff4-4113-8e6a-32f03be5c734"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.956220 4799 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a30589ef-eff4-4113-8e6a-32f03be5c734-lock\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.956544 4799 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.956584 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 24 07:09:41 crc kubenswrapper[4799]: I1124 07:09:41.956597 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckrjg\" (UniqueName: \"kubernetes.io/projected/a30589ef-eff4-4113-8e6a-32f03be5c734-kube-api-access-ckrjg\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.010119 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.087917 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.284873 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kgrrs_a7742c81-45e0-4ef5-aceb-7a631bf9cc37/ovs-vswitchd/0.log" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.286466 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.294168 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4nd7\" (UniqueName: \"kubernetes.io/projected/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-kube-api-access-r4nd7\") pod \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.294285 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-scripts\") pod \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.294332 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-lib\") pod \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.294356 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-log\") pod \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.294406 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-etc-ovs\") pod \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.294427 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-run\") pod \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\" (UID: \"a7742c81-45e0-4ef5-aceb-7a631bf9cc37\") " Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.294539 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-log" (OuterVolumeSpecName: "var-log") pod "a7742c81-45e0-4ef5-aceb-7a631bf9cc37" (UID: "a7742c81-45e0-4ef5-aceb-7a631bf9cc37"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.294633 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-lib" (OuterVolumeSpecName: "var-lib") pod "a7742c81-45e0-4ef5-aceb-7a631bf9cc37" (UID: "a7742c81-45e0-4ef5-aceb-7a631bf9cc37"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.294670 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "a7742c81-45e0-4ef5-aceb-7a631bf9cc37" (UID: "a7742c81-45e0-4ef5-aceb-7a631bf9cc37"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.294701 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-run" (OuterVolumeSpecName: "var-run") pod "a7742c81-45e0-4ef5-aceb-7a631bf9cc37" (UID: "a7742c81-45e0-4ef5-aceb-7a631bf9cc37"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.295137 4799 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-lib\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.295167 4799 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-log\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.295176 4799 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-etc-ovs\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.295185 4799 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.295664 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-scripts" (OuterVolumeSpecName: "scripts") pod "a7742c81-45e0-4ef5-aceb-7a631bf9cc37" (UID: "a7742c81-45e0-4ef5-aceb-7a631bf9cc37"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.299006 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-kube-api-access-r4nd7" (OuterVolumeSpecName: "kube-api-access-r4nd7") pod "a7742c81-45e0-4ef5-aceb-7a631bf9cc37" (UID: "a7742c81-45e0-4ef5-aceb-7a631bf9cc37"). InnerVolumeSpecName "kube-api-access-r4nd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.397501 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.397542 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4nd7\" (UniqueName: \"kubernetes.io/projected/a7742c81-45e0-4ef5-aceb-7a631bf9cc37-kube-api-access-r4nd7\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.837919 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kgrrs_a7742c81-45e0-4ef5-aceb-7a631bf9cc37/ovs-vswitchd/0.log" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.839191 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kgrrs" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.839310 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.841043 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kgrrs" event={"ID":"a7742c81-45e0-4ef5-aceb-7a631bf9cc37","Type":"ContainerDied","Data":"d5dbdbfdad4349f320febaff6ac846a95686336f646a6f63854dac9181e3cd33"} Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.841120 4799 scope.go:117] "RemoveContainer" containerID="16b698c68fea2099ea2f0501fd7815b16ff9ac5a82ac85760e0bf74dd244663f" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.891264 4799 scope.go:117] "RemoveContainer" containerID="343d36adb241bb4bfeff9c68837ce257de16c2be6a63747d5d86718c1186eb81" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.892417 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-kgrrs"] Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.909174 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-kgrrs"] Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.921083 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.926759 4799 scope.go:117] "RemoveContainer" containerID="8bb8a5532f39845617e4497e4831243eb9b594226daca424f0f2fa01515b4c8b" Nov 24 07:09:42 crc kubenswrapper[4799]: I1124 07:09:42.927335 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Nov 24 07:09:43 crc kubenswrapper[4799]: I1124 07:09:43.638820 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" path="/var/lib/kubelet/pods/a30589ef-eff4-4113-8e6a-32f03be5c734/volumes" Nov 24 07:09:43 crc kubenswrapper[4799]: I1124 07:09:43.641603 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" path="/var/lib/kubelet/pods/a7742c81-45e0-4ef5-aceb-7a631bf9cc37/volumes" Nov 24 07:09:43 crc kubenswrapper[4799]: I1124 07:09:43.951135 4799 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod76771b43-063e-4bc4-a366-3a13f27e3124"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod76771b43-063e-4bc4-a366-3a13f27e3124] : Timed out while waiting for systemd to remove kubepods-besteffort-pod76771b43_063e_4bc4_a366_3a13f27e3124.slice" Nov 24 07:09:43 crc kubenswrapper[4799]: E1124 07:09:43.951205 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod76771b43-063e-4bc4-a366-3a13f27e3124] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod76771b43-063e-4bc4-a366-3a13f27e3124] : Timed out while waiting for systemd to remove kubepods-besteffort-pod76771b43_063e_4bc4_a366_3a13f27e3124.slice" pod="openstack/openstackclient" podUID="76771b43-063e-4bc4-a366-3a13f27e3124" Nov 24 07:09:44 crc kubenswrapper[4799]: I1124 07:09:44.865499 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.665347 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement2704-account-delete-ck9t2" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.671886 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell000c8-account-delete-n5mhs" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.704055 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nkk7\" (UniqueName: \"kubernetes.io/projected/b6f22313-6d9a-4b35-b892-695903d354b1-kube-api-access-8nkk7\") pod \"b6f22313-6d9a-4b35-b892-695903d354b1\" (UID: \"b6f22313-6d9a-4b35-b892-695903d354b1\") " Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.704159 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts\") pod \"b6f22313-6d9a-4b35-b892-695903d354b1\" (UID: \"b6f22313-6d9a-4b35-b892-695903d354b1\") " Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.704187 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5b78\" (UniqueName: \"kubernetes.io/projected/87439e47-6c84-4467-a49b-03ed6e86a9eb-kube-api-access-v5b78\") pod \"87439e47-6c84-4467-a49b-03ed6e86a9eb\" (UID: \"87439e47-6c84-4467-a49b-03ed6e86a9eb\") " Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.704290 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts\") pod \"87439e47-6c84-4467-a49b-03ed6e86a9eb\" (UID: \"87439e47-6c84-4467-a49b-03ed6e86a9eb\") " Nov 24 07:09:47 crc kubenswrapper[4799]: E1124 07:09:47.704761 4799 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 24 07:09:47 crc kubenswrapper[4799]: E1124 07:09:47.704859 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts podName:d13b66c3-1e31-418e-a0ff-0d1bf7a5980a nodeName:}" failed. No retries permitted until 2025-11-24 07:10:19.704826258 +0000 UTC m=+1365.360808722 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts") pod "barbican44f2-account-delete-2hm9q" (UID: "d13b66c3-1e31-418e-a0ff-0d1bf7a5980a") : configmap "openstack-scripts" not found Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.705563 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b6f22313-6d9a-4b35-b892-695903d354b1" (UID: "b6f22313-6d9a-4b35-b892-695903d354b1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.707194 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "87439e47-6c84-4467-a49b-03ed6e86a9eb" (UID: "87439e47-6c84-4467-a49b-03ed6e86a9eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.712808 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87439e47-6c84-4467-a49b-03ed6e86a9eb-kube-api-access-v5b78" (OuterVolumeSpecName: "kube-api-access-v5b78") pod "87439e47-6c84-4467-a49b-03ed6e86a9eb" (UID: "87439e47-6c84-4467-a49b-03ed6e86a9eb"). InnerVolumeSpecName "kube-api-access-v5b78". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.712881 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6f22313-6d9a-4b35-b892-695903d354b1-kube-api-access-8nkk7" (OuterVolumeSpecName: "kube-api-access-8nkk7") pod "b6f22313-6d9a-4b35-b892-695903d354b1" (UID: "b6f22313-6d9a-4b35-b892-695903d354b1"). InnerVolumeSpecName "kube-api-access-8nkk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.806031 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87439e47-6c84-4467-a49b-03ed6e86a9eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.806099 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nkk7\" (UniqueName: \"kubernetes.io/projected/b6f22313-6d9a-4b35-b892-695903d354b1-kube-api-access-8nkk7\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.806123 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f22313-6d9a-4b35-b892-695903d354b1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.806141 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5b78\" (UniqueName: \"kubernetes.io/projected/87439e47-6c84-4467-a49b-03ed6e86a9eb-kube-api-access-v5b78\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.907227 4799 generic.go:334] "Generic (PLEG): container finished" podID="b6f22313-6d9a-4b35-b892-695903d354b1" containerID="fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61" exitCode=137 Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.907371 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement2704-account-delete-ck9t2" event={"ID":"b6f22313-6d9a-4b35-b892-695903d354b1","Type":"ContainerDied","Data":"fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61"} Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.907417 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement2704-account-delete-ck9t2" event={"ID":"b6f22313-6d9a-4b35-b892-695903d354b1","Type":"ContainerDied","Data":"9a2d2c63d77347f5059d85279c8ba9c130ff9e52541316e83e9fd29653b25b04"} Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.907448 4799 scope.go:117] "RemoveContainer" containerID="fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.907642 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement2704-account-delete-ck9t2" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.912695 4799 generic.go:334] "Generic (PLEG): container finished" podID="87439e47-6c84-4467-a49b-03ed6e86a9eb" containerID="44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6" exitCode=137 Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.912755 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell000c8-account-delete-n5mhs" event={"ID":"87439e47-6c84-4467-a49b-03ed6e86a9eb","Type":"ContainerDied","Data":"44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6"} Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.912796 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell000c8-account-delete-n5mhs" event={"ID":"87439e47-6c84-4467-a49b-03ed6e86a9eb","Type":"ContainerDied","Data":"f56d12bad72bbd48e989299f0bd988a09be84a662c5b6cba9efd1892f6217316"} Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.912893 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell000c8-account-delete-n5mhs" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.946976 4799 scope.go:117] "RemoveContainer" containerID="fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61" Nov 24 07:09:47 crc kubenswrapper[4799]: E1124 07:09:47.947735 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61\": container with ID starting with fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61 not found: ID does not exist" containerID="fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.947836 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61"} err="failed to get container status \"fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61\": rpc error: code = NotFound desc = could not find container \"fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61\": container with ID starting with fcbb3dcf28a26bd5f2349c01e2f4b9c9c46e69722226da2899883991d5087f61 not found: ID does not exist" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.947944 4799 scope.go:117] "RemoveContainer" containerID="44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.973488 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell000c8-account-delete-n5mhs"] Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.984720 4799 scope.go:117] "RemoveContainer" containerID="44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.984716 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell000c8-account-delete-n5mhs"] Nov 24 07:09:47 crc kubenswrapper[4799]: E1124 07:09:47.985620 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6\": container with ID starting with 44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6 not found: ID does not exist" containerID="44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.985712 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6"} err="failed to get container status \"44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6\": rpc error: code = NotFound desc = could not find container \"44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6\": container with ID starting with 44babb1b2f5ef8174a77f11246982865f16ec83c2f85da4fcfc8d706face11a6 not found: ID does not exist" Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.992002 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement2704-account-delete-ck9t2"] Nov 24 07:09:47 crc kubenswrapper[4799]: I1124 07:09:47.996485 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement2704-account-delete-ck9t2"] Nov 24 07:09:49 crc kubenswrapper[4799]: I1124 07:09:49.641065 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87439e47-6c84-4467-a49b-03ed6e86a9eb" path="/var/lib/kubelet/pods/87439e47-6c84-4467-a49b-03ed6e86a9eb/volumes" Nov 24 07:09:49 crc kubenswrapper[4799]: I1124 07:09:49.642450 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6f22313-6d9a-4b35-b892-695903d354b1" path="/var/lib/kubelet/pods/b6f22313-6d9a-4b35-b892-695903d354b1/volumes" Nov 24 07:09:50 crc kubenswrapper[4799]: E1124 07:09:50.371005 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd13b66c3_1e31_418e_a0ff_0d1bf7a5980a.slice/crio-16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090.scope\": RecentStats: unable to find data in memory cache]" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.400334 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.400382 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.400422 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.400985 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5003777d46acb4b46479f05a79e8d64bd64dee47094406d45a1866a4471e6cd1"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.401042 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://5003777d46acb4b46479f05a79e8d64bd64dee47094406d45a1866a4471e6cd1" gracePeriod=600 Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.732553 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican44f2-account-delete-2hm9q" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.772539 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts\") pod \"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a\" (UID: \"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a\") " Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.772996 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4cdf\" (UniqueName: \"kubernetes.io/projected/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-kube-api-access-c4cdf\") pod \"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a\" (UID: \"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a\") " Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.774789 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d13b66c3-1e31-418e-a0ff-0d1bf7a5980a" (UID: "d13b66c3-1e31-418e-a0ff-0d1bf7a5980a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.786902 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-kube-api-access-c4cdf" (OuterVolumeSpecName: "kube-api-access-c4cdf") pod "d13b66c3-1e31-418e-a0ff-0d1bf7a5980a" (UID: "d13b66c3-1e31-418e-a0ff-0d1bf7a5980a"). InnerVolumeSpecName "kube-api-access-c4cdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.874454 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4cdf\" (UniqueName: \"kubernetes.io/projected/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-kube-api-access-c4cdf\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.874804 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.959607 4799 generic.go:334] "Generic (PLEG): container finished" podID="d13b66c3-1e31-418e-a0ff-0d1bf7a5980a" containerID="16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090" exitCode=137 Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.959683 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican44f2-account-delete-2hm9q" event={"ID":"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a","Type":"ContainerDied","Data":"16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090"} Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.959711 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican44f2-account-delete-2hm9q" event={"ID":"d13b66c3-1e31-418e-a0ff-0d1bf7a5980a","Type":"ContainerDied","Data":"5b4890d36b23a1d16cc862079b4d9b8b999f8afb812820c384015d6702072fbf"} Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.959729 4799 scope.go:117] "RemoveContainer" containerID="16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.959747 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican44f2-account-delete-2hm9q" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.963318 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="5003777d46acb4b46479f05a79e8d64bd64dee47094406d45a1866a4471e6cd1" exitCode=0 Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.963345 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"5003777d46acb4b46479f05a79e8d64bd64dee47094406d45a1866a4471e6cd1"} Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.963361 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6"} Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.994916 4799 scope.go:117] "RemoveContainer" containerID="16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090" Nov 24 07:09:50 crc kubenswrapper[4799]: E1124 07:09:50.995698 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090\": container with ID starting with 16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090 not found: ID does not exist" containerID="16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.995791 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090"} err="failed to get container status \"16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090\": rpc error: code = NotFound desc = could not find container \"16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090\": container with ID starting with 16d722713423d67a6021cb30abfc4fd1da626929a5d9db6e5ee096177c119090 not found: ID does not exist" Nov 24 07:09:50 crc kubenswrapper[4799]: I1124 07:09:50.995843 4799 scope.go:117] "RemoveContainer" containerID="9829ac623e13e600dbb23e96d6ce6d0e396075d576f57770ed87771c2d8517cf" Nov 24 07:09:51 crc kubenswrapper[4799]: I1124 07:09:51.016944 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican44f2-account-delete-2hm9q"] Nov 24 07:09:51 crc kubenswrapper[4799]: I1124 07:09:51.025023 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican44f2-account-delete-2hm9q"] Nov 24 07:09:51 crc kubenswrapper[4799]: I1124 07:09:51.641361 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d13b66c3-1e31-418e-a0ff-0d1bf7a5980a" path="/var/lib/kubelet/pods/d13b66c3-1e31-418e-a0ff-0d1bf7a5980a/volumes" Nov 24 07:10:50 crc kubenswrapper[4799]: I1124 07:10:50.023254 4799 scope.go:117] "RemoveContainer" containerID="e4394c9753dd8146da5af402052bb93b4dc93f3348696e0f8ee006cd1df32c5c" Nov 24 07:10:50 crc kubenswrapper[4799]: I1124 07:10:50.055791 4799 scope.go:117] "RemoveContainer" containerID="c54fcecc5a5d26755e67a7a6d3e539af8f51f5fc6dd58f3a088edd47fd54d01d" Nov 24 07:10:50 crc kubenswrapper[4799]: I1124 07:10:50.095542 4799 scope.go:117] "RemoveContainer" containerID="6c441a620e5b48c41c3cf1709a813e69634cba1f9b860df79644b13ccfc40292" Nov 24 07:10:50 crc kubenswrapper[4799]: I1124 07:10:50.142787 4799 scope.go:117] "RemoveContainer" containerID="fd71cd5d765ff4cf61393c1642ef7e752e23da9d8a6512420626cd2e2713c9a3" Nov 24 07:10:50 crc kubenswrapper[4799]: I1124 07:10:50.179530 4799 scope.go:117] "RemoveContainer" containerID="ffedbe819747f9260bf81488c2e8b318f2f357ada47ae430c3886e5eda43a1fd" Nov 24 07:10:50 crc kubenswrapper[4799]: I1124 07:10:50.210030 4799 scope.go:117] "RemoveContainer" containerID="458ae0656978a384ca17474f208a13aed37fa1be14fa7cdbd76f5ef2136e7ccc" Nov 24 07:10:50 crc kubenswrapper[4799]: I1124 07:10:50.242733 4799 scope.go:117] "RemoveContainer" containerID="e8c89665cca133cc14d7182c955b9f158d183661351a87c4fa15352e6936b60e" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.401004 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.403009 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.479733 4799 scope.go:117] "RemoveContainer" containerID="adcebf3d8165db5c2482742ffe61dd12640c95e10c666abdf160cac3dd0d6b08" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.517939 4799 scope.go:117] "RemoveContainer" containerID="a21f45844786ca81306aa2947d5cce7e325c670a6b991864ca7472ab29d8c0b9" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.566246 4799 scope.go:117] "RemoveContainer" containerID="3328a1af0097e404c16240f7713c0a9b7963f642d250b69998ef457311035c34" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.599921 4799 scope.go:117] "RemoveContainer" containerID="9b90579a000b03e6467dabd651cbce2abb14ef949a00cfdd47d7596c85645587" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.631712 4799 scope.go:117] "RemoveContainer" containerID="479a19c5d4397919f77e002bce98bc8cfe8b88d861166ad64dbe89337e0e2126" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.662911 4799 scope.go:117] "RemoveContainer" containerID="ae5034ddd416e340224878800975aaca1a7d4c87688395fae504d9fb486b6005" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.699407 4799 scope.go:117] "RemoveContainer" containerID="3a79a583f1fe1d2c0c47fb1685562489608c16c24f320f11e04b10758e7e52b0" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.724620 4799 scope.go:117] "RemoveContainer" containerID="815111e7fbfc05e4da9663fc01168d0881dc15d3f9b8fe7ec1ce5a3f38cc6d40" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.748952 4799 scope.go:117] "RemoveContainer" containerID="34b20cfd09f348a8c1c28edcf64d822d884d0adade4ab63b86fe2ea2d6edf26d" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.775211 4799 scope.go:117] "RemoveContainer" containerID="6200c3568ecb120918ba43f8e8a656d9d1acb4db6b319da813f226c1af7e3025" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.801628 4799 scope.go:117] "RemoveContainer" containerID="dec54396850ed098f8d988491f08fcf315975d08d63a8517e087677c141683a7" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.823632 4799 scope.go:117] "RemoveContainer" containerID="c01da2c24a6c25ef981a4bf81e1ef3766e00137d2cf3e37995a2696e9e28633c" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.851630 4799 scope.go:117] "RemoveContainer" containerID="de17e7e1d402f63433a3438f6c8d6c44f1b24885c6c84bac90d88ef38ef8f5f7" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.873757 4799 scope.go:117] "RemoveContainer" containerID="a412226a89e3fd5d662ed68efbfbea22971efde4e4219cd6b56cf841e130fc92" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.889984 4799 scope.go:117] "RemoveContainer" containerID="c8e47fdf47fa1c6b2fdcb1d77c84b445e782cfd00343ab8b1883a72bed9feebd" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.908948 4799 scope.go:117] "RemoveContainer" containerID="2440fe858d8868b1c4904a513e11ec3d1af8a9378f9ca998eb5b15957accc132" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.934332 4799 scope.go:117] "RemoveContainer" containerID="bb57102415efc322a8bcd1bc7aa2a2615715511ee00dafa287e4f8bb115a5c1a" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.965606 4799 scope.go:117] "RemoveContainer" containerID="175f63c46417051e3c6f78ba7dd59612a708ca3b330b37ba4144be88999e43af" Nov 24 07:11:50 crc kubenswrapper[4799]: I1124 07:11:50.999254 4799 scope.go:117] "RemoveContainer" containerID="57ac5adb075544c22ae8aa32fb5357e1f72bc9a58d99ba17c8bb148b66dade15" Nov 24 07:11:51 crc kubenswrapper[4799]: I1124 07:11:51.026501 4799 scope.go:117] "RemoveContainer" containerID="2f8c8bdc7ce403fb12cce9fdf8de873b21760d472a20b313828bee7d09f25e41" Nov 24 07:11:51 crc kubenswrapper[4799]: I1124 07:11:51.076140 4799 scope.go:117] "RemoveContainer" containerID="6048bb73c3893b671ede7f9dc617529b1d172e7ca9f7861936ac9ca8e5189813" Nov 24 07:11:51 crc kubenswrapper[4799]: I1124 07:11:51.104090 4799 scope.go:117] "RemoveContainer" containerID="cdd7465ba59afe1601f7bbb0d441c7457c2da76592769f351a34b40d11ac5f9a" Nov 24 07:11:51 crc kubenswrapper[4799]: I1124 07:11:51.126406 4799 scope.go:117] "RemoveContainer" containerID="7a67d14359667b49d385e2378597910f70c85b133bae02d6c75678b330114be2" Nov 24 07:11:51 crc kubenswrapper[4799]: I1124 07:11:51.172381 4799 scope.go:117] "RemoveContainer" containerID="977a35126c8112db8b3dddf63c96480f5494a7696b5982d214d42310a46c7b83" Nov 24 07:11:51 crc kubenswrapper[4799]: I1124 07:11:51.198679 4799 scope.go:117] "RemoveContainer" containerID="6d52bb110c0d262b0946e49bcf01652ba0241c80b9cad0add83df423d3294499" Nov 24 07:11:51 crc kubenswrapper[4799]: I1124 07:11:51.216406 4799 scope.go:117] "RemoveContainer" containerID="8ac4f0ccf8167b086abf61aae9df350c81a1408a802be79619c6747aa2cc47af" Nov 24 07:11:51 crc kubenswrapper[4799]: I1124 07:11:51.233932 4799 scope.go:117] "RemoveContainer" containerID="22c9751adfda0a99f240920e420a888061186c08a987e917be24c8bc2d4dc9d3" Nov 24 07:11:51 crc kubenswrapper[4799]: I1124 07:11:51.253945 4799 scope.go:117] "RemoveContainer" containerID="a14fbfd0d90f8347c1f26eea42fb1a046ab2583fb22623e49ee3a3ffd3956b0c" Nov 24 07:11:51 crc kubenswrapper[4799]: I1124 07:11:51.297391 4799 scope.go:117] "RemoveContainer" containerID="5e46da699973a5e22e83b9aaac7e0e2ff7525ee80f32ff63ebc1fc46a7e36544" Nov 24 07:11:51 crc kubenswrapper[4799]: I1124 07:11:51.326951 4799 scope.go:117] "RemoveContainer" containerID="f9312779e4995c99fe48adfc65b9f134a79e29c78cf5915c12e2453b6490a6cd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.710437 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f22ng"] Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.711595 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a50ddac-d55d-4e99-8dea-8519b8d732d6" containerName="neutron-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.711616 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a50ddac-d55d-4e99-8dea-8519b8d732d6" containerName="neutron-api" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.711635 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" containerName="ovsdbserver-sb" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.711647 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" containerName="ovsdbserver-sb" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.711677 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server-init" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.711690 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server-init" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.711725 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-replicator" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.711737 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-replicator" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.711754 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.711766 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.711788 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerName="barbican-api-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.711800 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerName="barbican-api-log" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.711819 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feaa8e4-537c-44d2-ae92-ba999eb1626c" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.711833 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feaa8e4-537c-44d2-ae92-ba999eb1626c" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.711885 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-server" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.711897 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-server" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.711916 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerName="nova-api-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.711930 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerName="nova-api-log" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.711950 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="168550d3-6597-41ff-adf4-d20f190c7a02" containerName="nova-cell1-conductor-conductor" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.711962 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="168550d3-6597-41ff-adf4-d20f190c7a02" containerName="nova-cell1-conductor-conductor" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.711986 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovs-vswitchd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.711998 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovs-vswitchd" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712019 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="swift-recon-cron" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712031 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="swift-recon-cron" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712052 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712065 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712078 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712090 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712108 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712121 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-log" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712138 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b639ad0e-8438-4a36-a776-ee53b6f0c080" containerName="proxy-server" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712150 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b639ad0e-8438-4a36-a776-ee53b6f0c080" containerName="proxy-server" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712179 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ae5dfe-6c96-4d2d-99cf-b792b139d1a1" containerName="nova-cell0-conductor-conductor" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712190 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ae5dfe-6c96-4d2d-99cf-b792b139d1a1" containerName="nova-cell0-conductor-conductor" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712214 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09082d4e-0623-40c5-b584-1eb58985cfc0" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712225 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="09082d4e-0623-40c5-b584-1eb58985cfc0" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712242 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-updater" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712256 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-updater" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712279 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a27166-5918-4ee7-8569-a9f58914d7be" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712290 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a27166-5918-4ee7-8569-a9f58914d7be" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712314 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-auditor" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712325 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-auditor" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712338 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f22313-6d9a-4b35-b892-695903d354b1" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712350 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f22313-6d9a-4b35-b892-695903d354b1" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712369 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63757cd4-2ed0-4423-b466-4dfac76e3ac4" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712381 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="63757cd4-2ed0-4423-b466-4dfac76e3ac4" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712396 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a50ddac-d55d-4e99-8dea-8519b8d732d6" containerName="neutron-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712409 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a50ddac-d55d-4e99-8dea-8519b8d732d6" containerName="neutron-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712431 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c986e5b7-9864-46ae-9c99-9072e9c6ce34" containerName="barbican-keystone-listener" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712442 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c986e5b7-9864-46ae-9c99-9072e9c6ce34" containerName="barbican-keystone-listener" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712461 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f643414c-72f0-4e7c-b804-20a56f3207ab" containerName="cinder-api-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712473 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f643414c-72f0-4e7c-b804-20a56f3207ab" containerName="cinder-api-log" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712488 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="844d53db-a457-4153-bdb4-d557ba34c099" containerName="galera" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712499 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="844d53db-a457-4153-bdb4-d557ba34c099" containerName="galera" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712513 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="ceilometer-central-agent" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712525 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="ceilometer-central-agent" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712542 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a036d8-ce56-4654-a782-15cd18730378" containerName="glance-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712554 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a036d8-ce56-4654-a782-15cd18730378" containerName="glance-log" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712566 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f86cb1-3f0e-4da8-82dd-e00522d4ceee" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712580 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f86cb1-3f0e-4da8-82dd-e00522d4ceee" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712596 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="844d53db-a457-4153-bdb4-d557ba34c099" containerName="mysql-bootstrap" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712608 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="844d53db-a457-4153-bdb4-d557ba34c099" containerName="mysql-bootstrap" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712626 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366b7548-2fca-4623-a45c-c1e0367ce93a" containerName="rabbitmq" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712638 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="366b7548-2fca-4623-a45c-c1e0367ce93a" containerName="rabbitmq" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712653 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" containerName="glance-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712665 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" containerName="glance-log" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712678 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce1d3095-b344-4e6c-b49a-d1de3a7260ee" containerName="dnsmasq-dns" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712689 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce1d3095-b344-4e6c-b49a-d1de3a7260ee" containerName="dnsmasq-dns" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712707 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9064d19-6806-4e78-beb4-cda924fe3010" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712719 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9064d19-6806-4e78-beb4-cda924fe3010" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712739 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-replicator" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712751 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-replicator" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712770 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f643414c-72f0-4e7c-b804-20a56f3207ab" containerName="cinder-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712782 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f643414c-72f0-4e7c-b804-20a56f3207ab" containerName="cinder-api" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712803 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-replicator" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712815 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-replicator" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712831 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7603a96c-19e2-48ce-9c5f-69eeacc8cd85" containerName="keystone-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712842 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7603a96c-19e2-48ce-9c5f-69eeacc8cd85" containerName="keystone-api" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712896 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" containerName="glance-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712908 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" containerName="glance-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712920 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96acf88b-3fe8-46ac-8393-37f720632dd6" containerName="rabbitmq" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712931 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="96acf88b-3fe8-46ac-8393-37f720632dd6" containerName="rabbitmq" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712945 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb94fabc-dfc3-40cc-9791-0272924692da" containerName="probe" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712957 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb94fabc-dfc3-40cc-9791-0272924692da" containerName="probe" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712969 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd44457-1418-4788-a029-82c54c917997" containerName="kube-state-metrics" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.712981 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd44457-1418-4788-a029-82c54c917997" containerName="kube-state-metrics" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.712996 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-auditor" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713007 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-auditor" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713028 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-reaper" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713040 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-reaper" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713057 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="897d0d90-bf27-4d5b-b136-863379656f34" containerName="placement-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713070 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="897d0d90-bf27-4d5b-b136-863379656f34" containerName="placement-log" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713088 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c986e5b7-9864-46ae-9c99-9072e9c6ce34" containerName="barbican-keystone-listener-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713099 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c986e5b7-9864-46ae-9c99-9072e9c6ce34" containerName="barbican-keystone-listener-log" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713114 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366b7548-2fca-4623-a45c-c1e0367ce93a" containerName="setup-container" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713125 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="366b7548-2fca-4623-a45c-c1e0367ce93a" containerName="setup-container" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713141 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-updater" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713152 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-updater" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713170 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce1d3095-b344-4e6c-b49a-d1de3a7260ee" containerName="init" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713182 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce1d3095-b344-4e6c-b49a-d1de3a7260ee" containerName="init" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713199 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="sg-core" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713212 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="sg-core" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713235 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997ffb9e-cc6a-4fca-91d9-eb17d711b135" containerName="galera" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713246 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="997ffb9e-cc6a-4fca-91d9-eb17d711b135" containerName="galera" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713268 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b639ad0e-8438-4a36-a776-ee53b6f0c080" containerName="proxy-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713281 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b639ad0e-8438-4a36-a776-ee53b6f0c080" containerName="proxy-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713302 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96acf88b-3fe8-46ac-8393-37f720632dd6" containerName="setup-container" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713315 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="96acf88b-3fe8-46ac-8393-37f720632dd6" containerName="setup-container" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713335 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ee9f9f-66be-467a-a982-9ede1e1b682c" containerName="barbican-worker-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713347 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ee9f9f-66be-467a-a982-9ede1e1b682c" containerName="barbican-worker-log" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713367 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0361baf-d0c3-430f-87e3-c011f869a2b3" containerName="memcached" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713379 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0361baf-d0c3-430f-87e3-c011f869a2b3" containerName="memcached" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713395 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="rsync" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713406 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="rsync" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713421 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerName="ovn-northd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713433 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerName="ovn-northd" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713445 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbfb60f9-6e71-4b22-9d74-cab607a00c20" containerName="ovsdbserver-nb" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713457 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbfb60f9-6e71-4b22-9d74-cab607a00c20" containerName="ovsdbserver-nb" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713471 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerName="barbican-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713482 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerName="barbican-api" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713500 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d55de08-d6b7-44f6-8e96-44fcd8999000" containerName="nova-scheduler-scheduler" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713512 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d55de08-d6b7-44f6-8e96-44fcd8999000" containerName="nova-scheduler-scheduler" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713526 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-metadata" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713537 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-metadata" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713556 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87439e47-6c84-4467-a49b-03ed6e86a9eb" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713568 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="87439e47-6c84-4467-a49b-03ed6e86a9eb" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713589 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd32bea6-7ad9-4182-bc59-5644efe24ce1" containerName="ovn-controller" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713601 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd32bea6-7ad9-4182-bc59-5644efe24ce1" containerName="ovn-controller" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713619 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb94fabc-dfc3-40cc-9791-0272924692da" containerName="cinder-scheduler" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713631 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb94fabc-dfc3-40cc-9791-0272924692da" containerName="cinder-scheduler" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713646 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbfb60f9-6e71-4b22-9d74-cab607a00c20" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713658 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbfb60f9-6e71-4b22-9d74-cab607a00c20" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713671 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="897d0d90-bf27-4d5b-b136-863379656f34" containerName="placement-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713683 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="897d0d90-bf27-4d5b-b136-863379656f34" containerName="placement-api" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713703 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad75f933-ab36-4c9d-857d-c2cc11702dd3" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713715 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad75f933-ab36-4c9d-857d-c2cc11702dd3" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713734 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerName="nova-api-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713745 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerName="nova-api-api" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713761 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997ffb9e-cc6a-4fca-91d9-eb17d711b135" containerName="mysql-bootstrap" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713773 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="997ffb9e-cc6a-4fca-91d9-eb17d711b135" containerName="mysql-bootstrap" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713790 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-server" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713802 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-server" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713816 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ee9f9f-66be-467a-a982-9ede1e1b682c" containerName="barbican-worker" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713828 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ee9f9f-66be-467a-a982-9ede1e1b682c" containerName="barbican-worker" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713843 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-server" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713874 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-server" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713895 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d13b66c3-1e31-418e-a0ff-0d1bf7a5980a" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713907 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d13b66c3-1e31-418e-a0ff-0d1bf7a5980a" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713929 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a036d8-ce56-4654-a782-15cd18730378" containerName="glance-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713941 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a036d8-ce56-4654-a782-15cd18730378" containerName="glance-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713960 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-auditor" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.713973 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-auditor" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.713992 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="proxy-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714004 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="proxy-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.714024 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="ceilometer-notification-agent" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714036 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="ceilometer-notification-agent" Nov 24 07:11:59 crc kubenswrapper[4799]: E1124 07:11:59.714056 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-expirer" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714067 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-expirer" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714298 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbfb60f9-6e71-4b22-9d74-cab607a00c20" containerName="ovsdbserver-nb" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714324 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-updater" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714340 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9a036d8-ce56-4654-a782-15cd18730378" containerName="glance-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714357 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="63757cd4-2ed0-4423-b466-4dfac76e3ac4" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714375 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="997ffb9e-cc6a-4fca-91d9-eb17d711b135" containerName="galera" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714394 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="09082d4e-0623-40c5-b584-1eb58985cfc0" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714412 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovsdb-server" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714424 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" containerName="ovsdbserver-sb" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714447 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-metadata" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714460 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f643414c-72f0-4e7c-b804-20a56f3207ab" containerName="cinder-api-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714474 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7603a96c-19e2-48ce-9c5f-69eeacc8cd85" containerName="keystone-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714494 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" containerName="glance-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714508 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-replicator" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714531 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="swift-recon-cron" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714573 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb94fabc-dfc3-40cc-9791-0272924692da" containerName="cinder-scheduler" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714589 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="87439e47-6c84-4467-a49b-03ed6e86a9eb" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714608 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="96acf88b-3fe8-46ac-8393-37f720632dd6" containerName="rabbitmq" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714621 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714641 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-auditor" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714660 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-server" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714681 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="897d0d90-bf27-4d5b-b136-863379656f34" containerName="placement-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714699 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0361baf-d0c3-430f-87e3-c011f869a2b3" containerName="memcached" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714716 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce1d3095-b344-4e6c-b49a-d1de3a7260ee" containerName="dnsmasq-dns" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714732 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a50ddac-d55d-4e99-8dea-8519b8d732d6" containerName="neutron-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714748 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3feaa8e4-537c-44d2-ae92-ba999eb1626c" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714767 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-server" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.714786 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerName="barbican-api-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.715109 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eb15cf1-e5d8-43d7-9ad3-07a1c26b0ace" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.715139 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-replicator" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718002 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b639ad0e-8438-4a36-a776-ee53b6f0c080" containerName="proxy-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718035 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f643414c-72f0-4e7c-b804-20a56f3207ab" containerName="cinder-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718051 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="66ae5dfe-6c96-4d2d-99cf-b792b139d1a1" containerName="nova-cell0-conductor-conductor" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718064 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="sg-core" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718083 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ee9f9f-66be-467a-a982-9ede1e1b682c" containerName="barbican-worker-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718102 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b639ad0e-8438-4a36-a776-ee53b6f0c080" containerName="proxy-server" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718124 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9064d19-6806-4e78-beb4-cda924fe3010" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718143 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7742c81-45e0-4ef5-aceb-7a631bf9cc37" containerName="ovs-vswitchd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718161 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b14e8a8-2e19-47fd-a440-ac8d0cafecaf" containerName="glance-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718178 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad75f933-ab36-4c9d-857d-c2cc11702dd3" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718200 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="168550d3-6597-41ff-adf4-d20f190c7a02" containerName="nova-cell1-conductor-conductor" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718224 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="proxy-httpd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718270 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6f22313-6d9a-4b35-b892-695903d354b1" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718288 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="rsync" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718308 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebfe77df-bbcd-412f-9703-78e23d93f410" containerName="barbican-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718321 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9a036d8-ce56-4654-a782-15cd18730378" containerName="glance-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718336 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="ceilometer-central-agent" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718351 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="844d53db-a457-4153-bdb4-d557ba34c099" containerName="galera" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718368 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-server" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718388 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-auditor" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718405 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbfb60f9-6e71-4b22-9d74-cab607a00c20" containerName="openstack-network-exporter" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718420 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f97d9a56-cb78-4da9-b832-0028b6e45895" containerName="ceilometer-notification-agent" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718438 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd44457-1418-4788-a029-82c54c917997" containerName="kube-state-metrics" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718450 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f86cb1-3f0e-4da8-82dd-e00522d4ceee" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718464 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb94fabc-dfc3-40cc-9791-0272924692da" containerName="probe" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718483 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="account-reaper" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718498 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c986e5b7-9864-46ae-9c99-9072e9c6ce34" containerName="barbican-keystone-listener" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718514 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerName="nova-api-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718529 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-updater" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718544 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4a27166-5918-4ee7-8569-a9f58914d7be" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718578 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd32bea6-7ad9-4182-bc59-5644efe24ce1" containerName="ovn-controller" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718592 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fe03a76-2eb9-4728-9396-9d2dc1e743af" containerName="nova-api-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718611 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6218522b-3df9-45b0-8edb-fb935a7f4109" containerName="nova-metadata-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718631 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa59428-ee60-4434-9245-3ff8fa55200e" containerName="ovn-northd" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718645 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-auditor" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718660 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d55de08-d6b7-44f6-8e96-44fcd8999000" containerName="nova-scheduler-scheduler" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718677 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="366b7548-2fca-4623-a45c-c1e0367ce93a" containerName="rabbitmq" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718690 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="897d0d90-bf27-4d5b-b136-863379656f34" containerName="placement-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718704 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a50ddac-d55d-4e99-8dea-8519b8d732d6" containerName="neutron-api" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718726 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="container-replicator" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718749 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30589ef-eff4-4113-8e6a-32f03be5c734" containerName="object-expirer" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718769 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d13b66c3-1e31-418e-a0ff-0d1bf7a5980a" containerName="mariadb-account-delete" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718789 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c986e5b7-9864-46ae-9c99-9072e9c6ce34" containerName="barbican-keystone-listener-log" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.718810 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ee9f9f-66be-467a-a982-9ede1e1b682c" containerName="barbican-worker" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.721109 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.734985 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f22ng"] Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.790462 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-catalog-content\") pod \"certified-operators-f22ng\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.790561 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-utilities\") pod \"certified-operators-f22ng\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.790688 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9tfl\" (UniqueName: \"kubernetes.io/projected/0fac202e-11e8-463a-91f6-83bc5bfecbc8-kube-api-access-l9tfl\") pod \"certified-operators-f22ng\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.891877 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-catalog-content\") pod \"certified-operators-f22ng\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.891968 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-utilities\") pod \"certified-operators-f22ng\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.892032 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9tfl\" (UniqueName: \"kubernetes.io/projected/0fac202e-11e8-463a-91f6-83bc5bfecbc8-kube-api-access-l9tfl\") pod \"certified-operators-f22ng\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.892706 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-catalog-content\") pod \"certified-operators-f22ng\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.892744 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-utilities\") pod \"certified-operators-f22ng\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:11:59 crc kubenswrapper[4799]: I1124 07:11:59.925183 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9tfl\" (UniqueName: \"kubernetes.io/projected/0fac202e-11e8-463a-91f6-83bc5bfecbc8-kube-api-access-l9tfl\") pod \"certified-operators-f22ng\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:12:00 crc kubenswrapper[4799]: I1124 07:12:00.055184 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:12:00 crc kubenswrapper[4799]: I1124 07:12:00.567797 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f22ng"] Nov 24 07:12:01 crc kubenswrapper[4799]: I1124 07:12:01.418534 4799 generic.go:334] "Generic (PLEG): container finished" podID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" containerID="8702a8d5ebf8b7569df270df3ca9ad76a0b0704435ad940a4c18f32fc0ce24fc" exitCode=0 Nov 24 07:12:01 crc kubenswrapper[4799]: I1124 07:12:01.418663 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f22ng" event={"ID":"0fac202e-11e8-463a-91f6-83bc5bfecbc8","Type":"ContainerDied","Data":"8702a8d5ebf8b7569df270df3ca9ad76a0b0704435ad940a4c18f32fc0ce24fc"} Nov 24 07:12:01 crc kubenswrapper[4799]: I1124 07:12:01.418926 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f22ng" event={"ID":"0fac202e-11e8-463a-91f6-83bc5bfecbc8","Type":"ContainerStarted","Data":"f751c76d027df566b41be1ea89431cda32739810386a7388c2fba5e6f3d3c2e7"} Nov 24 07:12:01 crc kubenswrapper[4799]: I1124 07:12:01.420764 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:12:02 crc kubenswrapper[4799]: I1124 07:12:02.430118 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f22ng" event={"ID":"0fac202e-11e8-463a-91f6-83bc5bfecbc8","Type":"ContainerStarted","Data":"57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5"} Nov 24 07:12:03 crc kubenswrapper[4799]: I1124 07:12:03.458744 4799 generic.go:334] "Generic (PLEG): container finished" podID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" containerID="57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5" exitCode=0 Nov 24 07:12:03 crc kubenswrapper[4799]: I1124 07:12:03.458814 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f22ng" event={"ID":"0fac202e-11e8-463a-91f6-83bc5bfecbc8","Type":"ContainerDied","Data":"57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5"} Nov 24 07:12:04 crc kubenswrapper[4799]: I1124 07:12:04.472914 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f22ng" event={"ID":"0fac202e-11e8-463a-91f6-83bc5bfecbc8","Type":"ContainerStarted","Data":"40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83"} Nov 24 07:12:04 crc kubenswrapper[4799]: I1124 07:12:04.508541 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f22ng" podStartSLOduration=3.039659558 podStartE2EDuration="5.508511868s" podCreationTimestamp="2025-11-24 07:11:59 +0000 UTC" firstStartedPulling="2025-11-24 07:12:01.42055629 +0000 UTC m=+1467.076538774" lastFinishedPulling="2025-11-24 07:12:03.8894086 +0000 UTC m=+1469.545391084" observedRunningTime="2025-11-24 07:12:04.501558971 +0000 UTC m=+1470.157541475" watchObservedRunningTime="2025-11-24 07:12:04.508511868 +0000 UTC m=+1470.164494422" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.072212 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-svtxm"] Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.075015 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.126126 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-utilities\") pod \"community-operators-svtxm\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.126198 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhhbb\" (UniqueName: \"kubernetes.io/projected/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-kube-api-access-jhhbb\") pod \"community-operators-svtxm\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.126320 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-catalog-content\") pod \"community-operators-svtxm\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.129651 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svtxm"] Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.227508 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-catalog-content\") pod \"community-operators-svtxm\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.227616 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-utilities\") pod \"community-operators-svtxm\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.227650 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhhbb\" (UniqueName: \"kubernetes.io/projected/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-kube-api-access-jhhbb\") pod \"community-operators-svtxm\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.228235 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-catalog-content\") pod \"community-operators-svtxm\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.228315 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-utilities\") pod \"community-operators-svtxm\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.253909 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhhbb\" (UniqueName: \"kubernetes.io/projected/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-kube-api-access-jhhbb\") pod \"community-operators-svtxm\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.429521 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:06 crc kubenswrapper[4799]: I1124 07:12:06.940348 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svtxm"] Nov 24 07:12:06 crc kubenswrapper[4799]: W1124 07:12:06.941684 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27803ec4_81dd_4afa_8416_ccc3bfb33dc6.slice/crio-35dedf97af463d3e9d88d7547420f998bebe77fb37da449fde12490350b6b1a2 WatchSource:0}: Error finding container 35dedf97af463d3e9d88d7547420f998bebe77fb37da449fde12490350b6b1a2: Status 404 returned error can't find the container with id 35dedf97af463d3e9d88d7547420f998bebe77fb37da449fde12490350b6b1a2 Nov 24 07:12:07 crc kubenswrapper[4799]: I1124 07:12:07.503359 4799 generic.go:334] "Generic (PLEG): container finished" podID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" containerID="221e4d55f781dcdf9b9bcf44a09a116411d765ac4a7eb3b9db15e750c7ca4db0" exitCode=0 Nov 24 07:12:07 crc kubenswrapper[4799]: I1124 07:12:07.503441 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svtxm" event={"ID":"27803ec4-81dd-4afa-8416-ccc3bfb33dc6","Type":"ContainerDied","Data":"221e4d55f781dcdf9b9bcf44a09a116411d765ac4a7eb3b9db15e750c7ca4db0"} Nov 24 07:12:07 crc kubenswrapper[4799]: I1124 07:12:07.503607 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svtxm" event={"ID":"27803ec4-81dd-4afa-8416-ccc3bfb33dc6","Type":"ContainerStarted","Data":"35dedf97af463d3e9d88d7547420f998bebe77fb37da449fde12490350b6b1a2"} Nov 24 07:12:08 crc kubenswrapper[4799]: I1124 07:12:08.513876 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svtxm" event={"ID":"27803ec4-81dd-4afa-8416-ccc3bfb33dc6","Type":"ContainerStarted","Data":"7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b"} Nov 24 07:12:09 crc kubenswrapper[4799]: I1124 07:12:09.526825 4799 generic.go:334] "Generic (PLEG): container finished" podID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" containerID="7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b" exitCode=0 Nov 24 07:12:09 crc kubenswrapper[4799]: I1124 07:12:09.526983 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svtxm" event={"ID":"27803ec4-81dd-4afa-8416-ccc3bfb33dc6","Type":"ContainerDied","Data":"7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b"} Nov 24 07:12:10 crc kubenswrapper[4799]: I1124 07:12:10.055902 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:12:10 crc kubenswrapper[4799]: I1124 07:12:10.056446 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:12:10 crc kubenswrapper[4799]: I1124 07:12:10.118363 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:12:10 crc kubenswrapper[4799]: I1124 07:12:10.544911 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svtxm" event={"ID":"27803ec4-81dd-4afa-8416-ccc3bfb33dc6","Type":"ContainerStarted","Data":"371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38"} Nov 24 07:12:10 crc kubenswrapper[4799]: I1124 07:12:10.597671 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-svtxm" podStartSLOduration=2.097550617 podStartE2EDuration="4.597647543s" podCreationTimestamp="2025-11-24 07:12:06 +0000 UTC" firstStartedPulling="2025-11-24 07:12:07.505808175 +0000 UTC m=+1473.161790649" lastFinishedPulling="2025-11-24 07:12:10.005905061 +0000 UTC m=+1475.661887575" observedRunningTime="2025-11-24 07:12:10.594190305 +0000 UTC m=+1476.250172799" watchObservedRunningTime="2025-11-24 07:12:10.597647543 +0000 UTC m=+1476.253630027" Nov 24 07:12:10 crc kubenswrapper[4799]: I1124 07:12:10.654202 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:12:12 crc kubenswrapper[4799]: I1124 07:12:12.466264 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f22ng"] Nov 24 07:12:13 crc kubenswrapper[4799]: I1124 07:12:13.569950 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f22ng" podUID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" containerName="registry-server" containerID="cri-o://40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83" gracePeriod=2 Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.056775 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.244151 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-catalog-content\") pod \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.244275 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9tfl\" (UniqueName: \"kubernetes.io/projected/0fac202e-11e8-463a-91f6-83bc5bfecbc8-kube-api-access-l9tfl\") pod \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.244324 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-utilities\") pod \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\" (UID: \"0fac202e-11e8-463a-91f6-83bc5bfecbc8\") " Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.245572 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-utilities" (OuterVolumeSpecName: "utilities") pod "0fac202e-11e8-463a-91f6-83bc5bfecbc8" (UID: "0fac202e-11e8-463a-91f6-83bc5bfecbc8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.253330 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fac202e-11e8-463a-91f6-83bc5bfecbc8-kube-api-access-l9tfl" (OuterVolumeSpecName: "kube-api-access-l9tfl") pod "0fac202e-11e8-463a-91f6-83bc5bfecbc8" (UID: "0fac202e-11e8-463a-91f6-83bc5bfecbc8"). InnerVolumeSpecName "kube-api-access-l9tfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.298432 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fac202e-11e8-463a-91f6-83bc5bfecbc8" (UID: "0fac202e-11e8-463a-91f6-83bc5bfecbc8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.345828 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.345919 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9tfl\" (UniqueName: \"kubernetes.io/projected/0fac202e-11e8-463a-91f6-83bc5bfecbc8-kube-api-access-l9tfl\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.345945 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fac202e-11e8-463a-91f6-83bc5bfecbc8-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.584287 4799 generic.go:334] "Generic (PLEG): container finished" podID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" containerID="40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83" exitCode=0 Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.584334 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f22ng" event={"ID":"0fac202e-11e8-463a-91f6-83bc5bfecbc8","Type":"ContainerDied","Data":"40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83"} Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.584364 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f22ng" event={"ID":"0fac202e-11e8-463a-91f6-83bc5bfecbc8","Type":"ContainerDied","Data":"f751c76d027df566b41be1ea89431cda32739810386a7388c2fba5e6f3d3c2e7"} Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.584387 4799 scope.go:117] "RemoveContainer" containerID="40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.584448 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f22ng" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.614607 4799 scope.go:117] "RemoveContainer" containerID="57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.641507 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f22ng"] Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.653985 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f22ng"] Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.662250 4799 scope.go:117] "RemoveContainer" containerID="8702a8d5ebf8b7569df270df3ca9ad76a0b0704435ad940a4c18f32fc0ce24fc" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.694517 4799 scope.go:117] "RemoveContainer" containerID="40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83" Nov 24 07:12:14 crc kubenswrapper[4799]: E1124 07:12:14.695748 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83\": container with ID starting with 40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83 not found: ID does not exist" containerID="40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.695806 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83"} err="failed to get container status \"40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83\": rpc error: code = NotFound desc = could not find container \"40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83\": container with ID starting with 40f0a07c06f6a32ab968d3275835a22c6cba588024d0423b4abdaedef875fa83 not found: ID does not exist" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.695839 4799 scope.go:117] "RemoveContainer" containerID="57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5" Nov 24 07:12:14 crc kubenswrapper[4799]: E1124 07:12:14.696282 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5\": container with ID starting with 57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5 not found: ID does not exist" containerID="57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.696327 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5"} err="failed to get container status \"57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5\": rpc error: code = NotFound desc = could not find container \"57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5\": container with ID starting with 57b0b29297dc84bd48543509c4e7d3fc468dad6d5021b5d6c46fe7bebb0cc0a5 not found: ID does not exist" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.696359 4799 scope.go:117] "RemoveContainer" containerID="8702a8d5ebf8b7569df270df3ca9ad76a0b0704435ad940a4c18f32fc0ce24fc" Nov 24 07:12:14 crc kubenswrapper[4799]: E1124 07:12:14.696899 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8702a8d5ebf8b7569df270df3ca9ad76a0b0704435ad940a4c18f32fc0ce24fc\": container with ID starting with 8702a8d5ebf8b7569df270df3ca9ad76a0b0704435ad940a4c18f32fc0ce24fc not found: ID does not exist" containerID="8702a8d5ebf8b7569df270df3ca9ad76a0b0704435ad940a4c18f32fc0ce24fc" Nov 24 07:12:14 crc kubenswrapper[4799]: I1124 07:12:14.696944 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8702a8d5ebf8b7569df270df3ca9ad76a0b0704435ad940a4c18f32fc0ce24fc"} err="failed to get container status \"8702a8d5ebf8b7569df270df3ca9ad76a0b0704435ad940a4c18f32fc0ce24fc\": rpc error: code = NotFound desc = could not find container \"8702a8d5ebf8b7569df270df3ca9ad76a0b0704435ad940a4c18f32fc0ce24fc\": container with ID starting with 8702a8d5ebf8b7569df270df3ca9ad76a0b0704435ad940a4c18f32fc0ce24fc not found: ID does not exist" Nov 24 07:12:15 crc kubenswrapper[4799]: I1124 07:12:15.645692 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" path="/var/lib/kubelet/pods/0fac202e-11e8-463a-91f6-83bc5bfecbc8/volumes" Nov 24 07:12:16 crc kubenswrapper[4799]: I1124 07:12:16.430487 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:16 crc kubenswrapper[4799]: I1124 07:12:16.430920 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:16 crc kubenswrapper[4799]: I1124 07:12:16.476501 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:16 crc kubenswrapper[4799]: I1124 07:12:16.660887 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:17 crc kubenswrapper[4799]: I1124 07:12:17.481913 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svtxm"] Nov 24 07:12:18 crc kubenswrapper[4799]: I1124 07:12:18.623602 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-svtxm" podUID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" containerName="registry-server" containerID="cri-o://371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38" gracePeriod=2 Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.049677 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.216168 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-catalog-content\") pod \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.216328 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhhbb\" (UniqueName: \"kubernetes.io/projected/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-kube-api-access-jhhbb\") pod \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.216389 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-utilities\") pod \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\" (UID: \"27803ec4-81dd-4afa-8416-ccc3bfb33dc6\") " Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.218100 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-utilities" (OuterVolumeSpecName: "utilities") pod "27803ec4-81dd-4afa-8416-ccc3bfb33dc6" (UID: "27803ec4-81dd-4afa-8416-ccc3bfb33dc6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.225406 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-kube-api-access-jhhbb" (OuterVolumeSpecName: "kube-api-access-jhhbb") pod "27803ec4-81dd-4afa-8416-ccc3bfb33dc6" (UID: "27803ec4-81dd-4afa-8416-ccc3bfb33dc6"). InnerVolumeSpecName "kube-api-access-jhhbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.318812 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.318879 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhhbb\" (UniqueName: \"kubernetes.io/projected/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-kube-api-access-jhhbb\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.640588 4799 generic.go:334] "Generic (PLEG): container finished" podID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" containerID="371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38" exitCode=0 Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.640808 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svtxm" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.690330 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svtxm" event={"ID":"27803ec4-81dd-4afa-8416-ccc3bfb33dc6","Type":"ContainerDied","Data":"371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38"} Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.690397 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svtxm" event={"ID":"27803ec4-81dd-4afa-8416-ccc3bfb33dc6","Type":"ContainerDied","Data":"35dedf97af463d3e9d88d7547420f998bebe77fb37da449fde12490350b6b1a2"} Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.690426 4799 scope.go:117] "RemoveContainer" containerID="371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.717174 4799 scope.go:117] "RemoveContainer" containerID="7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.735587 4799 scope.go:117] "RemoveContainer" containerID="221e4d55f781dcdf9b9bcf44a09a116411d765ac4a7eb3b9db15e750c7ca4db0" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.800679 4799 scope.go:117] "RemoveContainer" containerID="371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38" Nov 24 07:12:19 crc kubenswrapper[4799]: E1124 07:12:19.801309 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38\": container with ID starting with 371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38 not found: ID does not exist" containerID="371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.801399 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38"} err="failed to get container status \"371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38\": rpc error: code = NotFound desc = could not find container \"371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38\": container with ID starting with 371adc559d7bc75baf9f0a5be9bbb368ddc3b6c22c1517fb2a6198f296e57d38 not found: ID does not exist" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.801527 4799 scope.go:117] "RemoveContainer" containerID="7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b" Nov 24 07:12:19 crc kubenswrapper[4799]: E1124 07:12:19.801888 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b\": container with ID starting with 7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b not found: ID does not exist" containerID="7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.801932 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b"} err="failed to get container status \"7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b\": rpc error: code = NotFound desc = could not find container \"7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b\": container with ID starting with 7d2a6e4bcfd37d472f4b303e72036ca143317e0596f7bd8626e36d9e3fcb083b not found: ID does not exist" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.801957 4799 scope.go:117] "RemoveContainer" containerID="221e4d55f781dcdf9b9bcf44a09a116411d765ac4a7eb3b9db15e750c7ca4db0" Nov 24 07:12:19 crc kubenswrapper[4799]: E1124 07:12:19.802381 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"221e4d55f781dcdf9b9bcf44a09a116411d765ac4a7eb3b9db15e750c7ca4db0\": container with ID starting with 221e4d55f781dcdf9b9bcf44a09a116411d765ac4a7eb3b9db15e750c7ca4db0 not found: ID does not exist" containerID="221e4d55f781dcdf9b9bcf44a09a116411d765ac4a7eb3b9db15e750c7ca4db0" Nov 24 07:12:19 crc kubenswrapper[4799]: I1124 07:12:19.802490 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"221e4d55f781dcdf9b9bcf44a09a116411d765ac4a7eb3b9db15e750c7ca4db0"} err="failed to get container status \"221e4d55f781dcdf9b9bcf44a09a116411d765ac4a7eb3b9db15e750c7ca4db0\": rpc error: code = NotFound desc = could not find container \"221e4d55f781dcdf9b9bcf44a09a116411d765ac4a7eb3b9db15e750c7ca4db0\": container with ID starting with 221e4d55f781dcdf9b9bcf44a09a116411d765ac4a7eb3b9db15e750c7ca4db0 not found: ID does not exist" Nov 24 07:12:20 crc kubenswrapper[4799]: I1124 07:12:20.049210 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27803ec4-81dd-4afa-8416-ccc3bfb33dc6" (UID: "27803ec4-81dd-4afa-8416-ccc3bfb33dc6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:20 crc kubenswrapper[4799]: I1124 07:12:20.130898 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27803ec4-81dd-4afa-8416-ccc3bfb33dc6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:20 crc kubenswrapper[4799]: I1124 07:12:20.302677 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svtxm"] Nov 24 07:12:20 crc kubenswrapper[4799]: I1124 07:12:20.313457 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-svtxm"] Nov 24 07:12:20 crc kubenswrapper[4799]: I1124 07:12:20.400464 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:12:20 crc kubenswrapper[4799]: I1124 07:12:20.400556 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:12:21 crc kubenswrapper[4799]: I1124 07:12:21.645742 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" path="/var/lib/kubelet/pods/27803ec4-81dd-4afa-8416-ccc3bfb33dc6/volumes" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.798248 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pcxg8"] Nov 24 07:12:40 crc kubenswrapper[4799]: E1124 07:12:40.799625 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" containerName="extract-content" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.799654 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" containerName="extract-content" Nov 24 07:12:40 crc kubenswrapper[4799]: E1124 07:12:40.799687 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" containerName="extract-utilities" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.799706 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" containerName="extract-utilities" Nov 24 07:12:40 crc kubenswrapper[4799]: E1124 07:12:40.799737 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" containerName="extract-utilities" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.799756 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" containerName="extract-utilities" Nov 24 07:12:40 crc kubenswrapper[4799]: E1124 07:12:40.799803 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" containerName="registry-server" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.799823 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" containerName="registry-server" Nov 24 07:12:40 crc kubenswrapper[4799]: E1124 07:12:40.799883 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" containerName="extract-content" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.799904 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" containerName="extract-content" Nov 24 07:12:40 crc kubenswrapper[4799]: E1124 07:12:40.799930 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" containerName="registry-server" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.799947 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" containerName="registry-server" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.800282 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="27803ec4-81dd-4afa-8416-ccc3bfb33dc6" containerName="registry-server" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.800317 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fac202e-11e8-463a-91f6-83bc5bfecbc8" containerName="registry-server" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.802972 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.808749 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-utilities\") pod \"redhat-operators-pcxg8\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.808994 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jxb8\" (UniqueName: \"kubernetes.io/projected/4c55bffc-8c52-4316-bb08-fa824e3c7035-kube-api-access-5jxb8\") pod \"redhat-operators-pcxg8\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.809055 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-catalog-content\") pod \"redhat-operators-pcxg8\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.810536 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pcxg8"] Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.911223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-utilities\") pod \"redhat-operators-pcxg8\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.911409 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jxb8\" (UniqueName: \"kubernetes.io/projected/4c55bffc-8c52-4316-bb08-fa824e3c7035-kube-api-access-5jxb8\") pod \"redhat-operators-pcxg8\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.911470 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-catalog-content\") pod \"redhat-operators-pcxg8\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.911826 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-utilities\") pod \"redhat-operators-pcxg8\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.912375 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-catalog-content\") pod \"redhat-operators-pcxg8\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:40 crc kubenswrapper[4799]: I1124 07:12:40.948391 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jxb8\" (UniqueName: \"kubernetes.io/projected/4c55bffc-8c52-4316-bb08-fa824e3c7035-kube-api-access-5jxb8\") pod \"redhat-operators-pcxg8\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:41 crc kubenswrapper[4799]: I1124 07:12:41.140655 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:41 crc kubenswrapper[4799]: I1124 07:12:41.356574 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pcxg8"] Nov 24 07:12:41 crc kubenswrapper[4799]: W1124 07:12:41.363746 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c55bffc_8c52_4316_bb08_fa824e3c7035.slice/crio-6ca7baf1b4cdd77956af26e660f9d072b322e7d9350c47413265b38c1b214c19 WatchSource:0}: Error finding container 6ca7baf1b4cdd77956af26e660f9d072b322e7d9350c47413265b38c1b214c19: Status 404 returned error can't find the container with id 6ca7baf1b4cdd77956af26e660f9d072b322e7d9350c47413265b38c1b214c19 Nov 24 07:12:41 crc kubenswrapper[4799]: I1124 07:12:41.881035 4799 generic.go:334] "Generic (PLEG): container finished" podID="4c55bffc-8c52-4316-bb08-fa824e3c7035" containerID="71dadafbe31ea4437745471040e9b0f9f7e62a57835081de3f482bc00ad1fd09" exitCode=0 Nov 24 07:12:41 crc kubenswrapper[4799]: I1124 07:12:41.881102 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcxg8" event={"ID":"4c55bffc-8c52-4316-bb08-fa824e3c7035","Type":"ContainerDied","Data":"71dadafbe31ea4437745471040e9b0f9f7e62a57835081de3f482bc00ad1fd09"} Nov 24 07:12:41 crc kubenswrapper[4799]: I1124 07:12:41.881174 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcxg8" event={"ID":"4c55bffc-8c52-4316-bb08-fa824e3c7035","Type":"ContainerStarted","Data":"6ca7baf1b4cdd77956af26e660f9d072b322e7d9350c47413265b38c1b214c19"} Nov 24 07:12:42 crc kubenswrapper[4799]: I1124 07:12:42.892328 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcxg8" event={"ID":"4c55bffc-8c52-4316-bb08-fa824e3c7035","Type":"ContainerStarted","Data":"c427a21f0f7b5834a3ef7d85692a06121737acabe11a6496bc83bfac2bdbeb59"} Nov 24 07:12:43 crc kubenswrapper[4799]: I1124 07:12:43.903366 4799 generic.go:334] "Generic (PLEG): container finished" podID="4c55bffc-8c52-4316-bb08-fa824e3c7035" containerID="c427a21f0f7b5834a3ef7d85692a06121737acabe11a6496bc83bfac2bdbeb59" exitCode=0 Nov 24 07:12:43 crc kubenswrapper[4799]: I1124 07:12:43.903449 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcxg8" event={"ID":"4c55bffc-8c52-4316-bb08-fa824e3c7035","Type":"ContainerDied","Data":"c427a21f0f7b5834a3ef7d85692a06121737acabe11a6496bc83bfac2bdbeb59"} Nov 24 07:12:44 crc kubenswrapper[4799]: I1124 07:12:44.913833 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcxg8" event={"ID":"4c55bffc-8c52-4316-bb08-fa824e3c7035","Type":"ContainerStarted","Data":"ae11b604a7a574041c4d809f168497761ba0f4b6c3b52982cf541a752e64575b"} Nov 24 07:12:44 crc kubenswrapper[4799]: I1124 07:12:44.935864 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pcxg8" podStartSLOduration=2.4878621450000002 podStartE2EDuration="4.935821729s" podCreationTimestamp="2025-11-24 07:12:40 +0000 UTC" firstStartedPulling="2025-11-24 07:12:41.88295117 +0000 UTC m=+1507.538933644" lastFinishedPulling="2025-11-24 07:12:44.330910764 +0000 UTC m=+1509.986893228" observedRunningTime="2025-11-24 07:12:44.932647168 +0000 UTC m=+1510.588629652" watchObservedRunningTime="2025-11-24 07:12:44.935821729 +0000 UTC m=+1510.591804213" Nov 24 07:12:50 crc kubenswrapper[4799]: I1124 07:12:50.400423 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:12:50 crc kubenswrapper[4799]: I1124 07:12:50.400832 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:12:50 crc kubenswrapper[4799]: I1124 07:12:50.400905 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:12:50 crc kubenswrapper[4799]: I1124 07:12:50.401595 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:12:50 crc kubenswrapper[4799]: I1124 07:12:50.401659 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" gracePeriod=600 Nov 24 07:12:50 crc kubenswrapper[4799]: E1124 07:12:50.530809 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:12:50 crc kubenswrapper[4799]: I1124 07:12:50.964271 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" exitCode=0 Nov 24 07:12:50 crc kubenswrapper[4799]: I1124 07:12:50.964332 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6"} Nov 24 07:12:50 crc kubenswrapper[4799]: I1124 07:12:50.964672 4799 scope.go:117] "RemoveContainer" containerID="5003777d46acb4b46479f05a79e8d64bd64dee47094406d45a1866a4471e6cd1" Nov 24 07:12:50 crc kubenswrapper[4799]: I1124 07:12:50.965508 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:12:50 crc kubenswrapper[4799]: E1124 07:12:50.966003 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.141820 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.141954 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.196641 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.637616 4799 scope.go:117] "RemoveContainer" containerID="bf402ab86034c98ced891cf3177f170ae03981592178d303ce73115272760cb3" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.661338 4799 scope.go:117] "RemoveContainer" containerID="482794ed63a9b329b5550fafc91a694f09f866c3013f86deef50df49ef6c173e" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.699472 4799 scope.go:117] "RemoveContainer" containerID="d3a425e1e004a0c21cb397f3983c5fe55824d01453fd599218ea20167f5095a8" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.731876 4799 scope.go:117] "RemoveContainer" containerID="40866b9dd65255e1c5dc3bd4fecb4975eab55bdd3723df4f474e46303e4b4a46" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.779371 4799 scope.go:117] "RemoveContainer" containerID="b1cbe7bfd60ea9dda2a4263ac59bad2dfc42f541a2a9046fdc4fb6a455efc23e" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.797753 4799 scope.go:117] "RemoveContainer" containerID="525cb566c1a708ef724e8e60124f3abc876a55149f6adbf23fede6052a7b4b2f" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.826532 4799 scope.go:117] "RemoveContainer" containerID="1bd8daa785503f4e2fcb67f8d5e0fa713f93ffd5b803f47263907f5420e39c50" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.852438 4799 scope.go:117] "RemoveContainer" containerID="caf4f5eb8dd9a9168446c28bcc88c19555451498f7a234dc9f24098b6462bfd0" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.922260 4799 scope.go:117] "RemoveContainer" containerID="1bae0067729d54c8ab8a099e80fb824001f7aaa34386e63f9c99d92e5d77407c" Nov 24 07:12:51 crc kubenswrapper[4799]: I1124 07:12:51.976089 4799 scope.go:117] "RemoveContainer" containerID="30959909b72adeb936fc238d713e097048621efd99d9e5661c63d6da5128b9e8" Nov 24 07:12:52 crc kubenswrapper[4799]: I1124 07:12:52.004803 4799 scope.go:117] "RemoveContainer" containerID="02ffa0bb3d383e7c692b859a1e3376a9f3f7187ae40a809812908dbb07a5603c" Nov 24 07:12:52 crc kubenswrapper[4799]: I1124 07:12:52.073480 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:52 crc kubenswrapper[4799]: I1124 07:12:52.120454 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pcxg8"] Nov 24 07:12:54 crc kubenswrapper[4799]: I1124 07:12:54.062055 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pcxg8" podUID="4c55bffc-8c52-4316-bb08-fa824e3c7035" containerName="registry-server" containerID="cri-o://ae11b604a7a574041c4d809f168497761ba0f4b6c3b52982cf541a752e64575b" gracePeriod=2 Nov 24 07:12:54 crc kubenswrapper[4799]: E1124 07:12:54.332197 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c55bffc_8c52_4316_bb08_fa824e3c7035.slice/crio-ae11b604a7a574041c4d809f168497761ba0f4b6c3b52982cf541a752e64575b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c55bffc_8c52_4316_bb08_fa824e3c7035.slice/crio-conmon-ae11b604a7a574041c4d809f168497761ba0f4b6c3b52982cf541a752e64575b.scope\": RecentStats: unable to find data in memory cache]" Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.072591 4799 generic.go:334] "Generic (PLEG): container finished" podID="4c55bffc-8c52-4316-bb08-fa824e3c7035" containerID="ae11b604a7a574041c4d809f168497761ba0f4b6c3b52982cf541a752e64575b" exitCode=0 Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.072642 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcxg8" event={"ID":"4c55bffc-8c52-4316-bb08-fa824e3c7035","Type":"ContainerDied","Data":"ae11b604a7a574041c4d809f168497761ba0f4b6c3b52982cf541a752e64575b"} Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.145882 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.318273 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-utilities\") pod \"4c55bffc-8c52-4316-bb08-fa824e3c7035\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.318358 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-catalog-content\") pod \"4c55bffc-8c52-4316-bb08-fa824e3c7035\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.318448 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jxb8\" (UniqueName: \"kubernetes.io/projected/4c55bffc-8c52-4316-bb08-fa824e3c7035-kube-api-access-5jxb8\") pod \"4c55bffc-8c52-4316-bb08-fa824e3c7035\" (UID: \"4c55bffc-8c52-4316-bb08-fa824e3c7035\") " Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.319550 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-utilities" (OuterVolumeSpecName: "utilities") pod "4c55bffc-8c52-4316-bb08-fa824e3c7035" (UID: "4c55bffc-8c52-4316-bb08-fa824e3c7035"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.323514 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c55bffc-8c52-4316-bb08-fa824e3c7035-kube-api-access-5jxb8" (OuterVolumeSpecName: "kube-api-access-5jxb8") pod "4c55bffc-8c52-4316-bb08-fa824e3c7035" (UID: "4c55bffc-8c52-4316-bb08-fa824e3c7035"). InnerVolumeSpecName "kube-api-access-5jxb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.419519 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.419546 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jxb8\" (UniqueName: \"kubernetes.io/projected/4c55bffc-8c52-4316-bb08-fa824e3c7035-kube-api-access-5jxb8\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.471254 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c55bffc-8c52-4316-bb08-fa824e3c7035" (UID: "4c55bffc-8c52-4316-bb08-fa824e3c7035"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:12:55 crc kubenswrapper[4799]: I1124 07:12:55.521207 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c55bffc-8c52-4316-bb08-fa824e3c7035-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:12:56 crc kubenswrapper[4799]: I1124 07:12:56.086036 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcxg8" event={"ID":"4c55bffc-8c52-4316-bb08-fa824e3c7035","Type":"ContainerDied","Data":"6ca7baf1b4cdd77956af26e660f9d072b322e7d9350c47413265b38c1b214c19"} Nov 24 07:12:56 crc kubenswrapper[4799]: I1124 07:12:56.086096 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcxg8" Nov 24 07:12:56 crc kubenswrapper[4799]: I1124 07:12:56.086127 4799 scope.go:117] "RemoveContainer" containerID="ae11b604a7a574041c4d809f168497761ba0f4b6c3b52982cf541a752e64575b" Nov 24 07:12:56 crc kubenswrapper[4799]: I1124 07:12:56.115946 4799 scope.go:117] "RemoveContainer" containerID="c427a21f0f7b5834a3ef7d85692a06121737acabe11a6496bc83bfac2bdbeb59" Nov 24 07:12:56 crc kubenswrapper[4799]: I1124 07:12:56.164589 4799 scope.go:117] "RemoveContainer" containerID="71dadafbe31ea4437745471040e9b0f9f7e62a57835081de3f482bc00ad1fd09" Nov 24 07:12:56 crc kubenswrapper[4799]: I1124 07:12:56.167082 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pcxg8"] Nov 24 07:12:56 crc kubenswrapper[4799]: I1124 07:12:56.179322 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pcxg8"] Nov 24 07:12:57 crc kubenswrapper[4799]: I1124 07:12:57.648924 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c55bffc-8c52-4316-bb08-fa824e3c7035" path="/var/lib/kubelet/pods/4c55bffc-8c52-4316-bb08-fa824e3c7035/volumes" Nov 24 07:13:02 crc kubenswrapper[4799]: I1124 07:13:02.628485 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:13:02 crc kubenswrapper[4799]: E1124 07:13:02.629296 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:13:15 crc kubenswrapper[4799]: I1124 07:13:15.639493 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:13:15 crc kubenswrapper[4799]: E1124 07:13:15.640544 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:13:30 crc kubenswrapper[4799]: I1124 07:13:30.628345 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:13:30 crc kubenswrapper[4799]: E1124 07:13:30.629350 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:13:44 crc kubenswrapper[4799]: I1124 07:13:44.629189 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:13:44 crc kubenswrapper[4799]: E1124 07:13:44.630382 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:13:52 crc kubenswrapper[4799]: I1124 07:13:52.180297 4799 scope.go:117] "RemoveContainer" containerID="c4a15beff8279fb5630bb3a06f94fc601d98cba69dc34d057239038e1cbe6e62" Nov 24 07:13:52 crc kubenswrapper[4799]: I1124 07:13:52.248334 4799 scope.go:117] "RemoveContainer" containerID="c138c8c4a165b11c836e1ac6661a13e3864e56da8aff5dba84d7403bc256e1bf" Nov 24 07:13:52 crc kubenswrapper[4799]: I1124 07:13:52.305184 4799 scope.go:117] "RemoveContainer" containerID="88399e1c222bc6dd7a643d5db0cd6a4e6b60cf5019ca0bbc1ca526de8ed139d3" Nov 24 07:13:52 crc kubenswrapper[4799]: I1124 07:13:52.346133 4799 scope.go:117] "RemoveContainer" containerID="4f037689b65a1da575275f13da9f49f1ca9ea45db68ab593237cee43fc871a9b" Nov 24 07:13:59 crc kubenswrapper[4799]: I1124 07:13:59.628910 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:13:59 crc kubenswrapper[4799]: E1124 07:13:59.630062 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:14:11 crc kubenswrapper[4799]: I1124 07:14:11.628838 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:14:11 crc kubenswrapper[4799]: E1124 07:14:11.629912 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:14:26 crc kubenswrapper[4799]: I1124 07:14:26.628735 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:14:26 crc kubenswrapper[4799]: E1124 07:14:26.630025 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:14:41 crc kubenswrapper[4799]: I1124 07:14:41.629217 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:14:41 crc kubenswrapper[4799]: E1124 07:14:41.630288 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:14:52 crc kubenswrapper[4799]: I1124 07:14:52.443718 4799 scope.go:117] "RemoveContainer" containerID="b59824ec9a02f2da4aecd1a2a4ddb873b0b15ff04f45160a866423c478cd5972" Nov 24 07:14:52 crc kubenswrapper[4799]: I1124 07:14:52.628089 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:14:52 crc kubenswrapper[4799]: E1124 07:14:52.628739 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.155894 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c"] Nov 24 07:15:00 crc kubenswrapper[4799]: E1124 07:15:00.157269 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c55bffc-8c52-4316-bb08-fa824e3c7035" containerName="extract-utilities" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.157302 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c55bffc-8c52-4316-bb08-fa824e3c7035" containerName="extract-utilities" Nov 24 07:15:00 crc kubenswrapper[4799]: E1124 07:15:00.157339 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c55bffc-8c52-4316-bb08-fa824e3c7035" containerName="extract-content" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.157357 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c55bffc-8c52-4316-bb08-fa824e3c7035" containerName="extract-content" Nov 24 07:15:00 crc kubenswrapper[4799]: E1124 07:15:00.157378 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c55bffc-8c52-4316-bb08-fa824e3c7035" containerName="registry-server" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.157402 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c55bffc-8c52-4316-bb08-fa824e3c7035" containerName="registry-server" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.157796 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c55bffc-8c52-4316-bb08-fa824e3c7035" containerName="registry-server" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.158765 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.160884 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.161923 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.165957 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c"] Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.347993 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-secret-volume\") pod \"collect-profiles-29399475-rqf9c\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.348712 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-config-volume\") pod \"collect-profiles-29399475-rqf9c\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.349025 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4kq6\" (UniqueName: \"kubernetes.io/projected/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-kube-api-access-w4kq6\") pod \"collect-profiles-29399475-rqf9c\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.450790 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4kq6\" (UniqueName: \"kubernetes.io/projected/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-kube-api-access-w4kq6\") pod \"collect-profiles-29399475-rqf9c\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.450973 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-secret-volume\") pod \"collect-profiles-29399475-rqf9c\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.451074 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-config-volume\") pod \"collect-profiles-29399475-rqf9c\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.452544 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-config-volume\") pod \"collect-profiles-29399475-rqf9c\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.478794 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-secret-volume\") pod \"collect-profiles-29399475-rqf9c\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.480538 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4kq6\" (UniqueName: \"kubernetes.io/projected/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-kube-api-access-w4kq6\") pod \"collect-profiles-29399475-rqf9c\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.510096 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:00 crc kubenswrapper[4799]: I1124 07:15:00.944108 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c"] Nov 24 07:15:01 crc kubenswrapper[4799]: I1124 07:15:01.272418 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" event={"ID":"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37","Type":"ContainerStarted","Data":"f9daf8b045e1ae6d8fb7ef3e6f13b5546f3d083e0d85b8c6d6cd6003673aeee6"} Nov 24 07:15:01 crc kubenswrapper[4799]: I1124 07:15:01.272751 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" event={"ID":"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37","Type":"ContainerStarted","Data":"446f3b4bfa966c0bb5b345a48271e47831af33853fae27bf3c07c340b3651807"} Nov 24 07:15:01 crc kubenswrapper[4799]: I1124 07:15:01.287366 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" podStartSLOduration=1.2873458740000001 podStartE2EDuration="1.287345874s" podCreationTimestamp="2025-11-24 07:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:15:01.284333986 +0000 UTC m=+1646.940316480" watchObservedRunningTime="2025-11-24 07:15:01.287345874 +0000 UTC m=+1646.943328348" Nov 24 07:15:02 crc kubenswrapper[4799]: I1124 07:15:02.284894 4799 generic.go:334] "Generic (PLEG): container finished" podID="54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37" containerID="f9daf8b045e1ae6d8fb7ef3e6f13b5546f3d083e0d85b8c6d6cd6003673aeee6" exitCode=0 Nov 24 07:15:02 crc kubenswrapper[4799]: I1124 07:15:02.284965 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" event={"ID":"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37","Type":"ContainerDied","Data":"f9daf8b045e1ae6d8fb7ef3e6f13b5546f3d083e0d85b8c6d6cd6003673aeee6"} Nov 24 07:15:03 crc kubenswrapper[4799]: I1124 07:15:03.609816 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:03 crc kubenswrapper[4799]: I1124 07:15:03.800743 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-config-volume\") pod \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " Nov 24 07:15:03 crc kubenswrapper[4799]: I1124 07:15:03.800977 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-secret-volume\") pod \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " Nov 24 07:15:03 crc kubenswrapper[4799]: I1124 07:15:03.801032 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4kq6\" (UniqueName: \"kubernetes.io/projected/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-kube-api-access-w4kq6\") pod \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\" (UID: \"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37\") " Nov 24 07:15:03 crc kubenswrapper[4799]: I1124 07:15:03.801768 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-config-volume" (OuterVolumeSpecName: "config-volume") pod "54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37" (UID: "54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:15:03 crc kubenswrapper[4799]: I1124 07:15:03.810294 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-kube-api-access-w4kq6" (OuterVolumeSpecName: "kube-api-access-w4kq6") pod "54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37" (UID: "54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37"). InnerVolumeSpecName "kube-api-access-w4kq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:15:03 crc kubenswrapper[4799]: I1124 07:15:03.811494 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37" (UID: "54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:15:03 crc kubenswrapper[4799]: I1124 07:15:03.902770 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:03 crc kubenswrapper[4799]: I1124 07:15:03.902823 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:03 crc kubenswrapper[4799]: I1124 07:15:03.902837 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4kq6\" (UniqueName: \"kubernetes.io/projected/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37-kube-api-access-w4kq6\") on node \"crc\" DevicePath \"\"" Nov 24 07:15:04 crc kubenswrapper[4799]: I1124 07:15:04.305458 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" event={"ID":"54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37","Type":"ContainerDied","Data":"446f3b4bfa966c0bb5b345a48271e47831af33853fae27bf3c07c340b3651807"} Nov 24 07:15:04 crc kubenswrapper[4799]: I1124 07:15:04.305516 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="446f3b4bfa966c0bb5b345a48271e47831af33853fae27bf3c07c340b3651807" Nov 24 07:15:04 crc kubenswrapper[4799]: I1124 07:15:04.305638 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c" Nov 24 07:15:07 crc kubenswrapper[4799]: I1124 07:15:07.628562 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:15:07 crc kubenswrapper[4799]: E1124 07:15:07.629299 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:15:20 crc kubenswrapper[4799]: I1124 07:15:20.629241 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:15:20 crc kubenswrapper[4799]: E1124 07:15:20.630488 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:15:32 crc kubenswrapper[4799]: I1124 07:15:32.628340 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:15:32 crc kubenswrapper[4799]: E1124 07:15:32.629382 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:15:45 crc kubenswrapper[4799]: I1124 07:15:45.639282 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:15:45 crc kubenswrapper[4799]: E1124 07:15:45.640422 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:15:52 crc kubenswrapper[4799]: I1124 07:15:52.556600 4799 scope.go:117] "RemoveContainer" containerID="0b036e7485bdbc7a10091975ed6eae68dfe9ad069cb89ee4aca3cd050c565d50" Nov 24 07:15:52 crc kubenswrapper[4799]: I1124 07:15:52.609049 4799 scope.go:117] "RemoveContainer" containerID="7d78517d5fccdbbf385d41ea0be66f81ad0ecb4529e90eaf0d139df4336fa2b1" Nov 24 07:15:52 crc kubenswrapper[4799]: I1124 07:15:52.648318 4799 scope.go:117] "RemoveContainer" containerID="d59b7428d4e280f7869415161a1fab26ded6505c40670360f51e2d5cf153d262" Nov 24 07:16:00 crc kubenswrapper[4799]: I1124 07:16:00.628932 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:16:00 crc kubenswrapper[4799]: E1124 07:16:00.629923 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:16:15 crc kubenswrapper[4799]: I1124 07:16:15.639862 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:16:15 crc kubenswrapper[4799]: E1124 07:16:15.641073 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:16:27 crc kubenswrapper[4799]: I1124 07:16:27.628602 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:16:27 crc kubenswrapper[4799]: E1124 07:16:27.629495 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:16:38 crc kubenswrapper[4799]: I1124 07:16:38.628261 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:16:38 crc kubenswrapper[4799]: E1124 07:16:38.628747 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:16:49 crc kubenswrapper[4799]: I1124 07:16:49.628579 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:16:49 crc kubenswrapper[4799]: E1124 07:16:49.629627 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:17:00 crc kubenswrapper[4799]: I1124 07:17:00.628742 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:17:00 crc kubenswrapper[4799]: E1124 07:17:00.629843 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:17:03 crc kubenswrapper[4799]: I1124 07:17:03.741811 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rhcsg"] Nov 24 07:17:03 crc kubenswrapper[4799]: E1124 07:17:03.742548 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37" containerName="collect-profiles" Nov 24 07:17:03 crc kubenswrapper[4799]: I1124 07:17:03.742566 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37" containerName="collect-profiles" Nov 24 07:17:03 crc kubenswrapper[4799]: I1124 07:17:03.743596 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37" containerName="collect-profiles" Nov 24 07:17:03 crc kubenswrapper[4799]: I1124 07:17:03.745046 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:03 crc kubenswrapper[4799]: I1124 07:17:03.817461 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rhcsg"] Nov 24 07:17:03 crc kubenswrapper[4799]: I1124 07:17:03.907059 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-catalog-content\") pod \"redhat-marketplace-rhcsg\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:03 crc kubenswrapper[4799]: I1124 07:17:03.907478 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-utilities\") pod \"redhat-marketplace-rhcsg\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:03 crc kubenswrapper[4799]: I1124 07:17:03.907564 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6jq6\" (UniqueName: \"kubernetes.io/projected/38de7098-5d03-4685-9ae3-0e96cd024915-kube-api-access-n6jq6\") pod \"redhat-marketplace-rhcsg\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:04 crc kubenswrapper[4799]: I1124 07:17:04.009257 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-catalog-content\") pod \"redhat-marketplace-rhcsg\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:04 crc kubenswrapper[4799]: I1124 07:17:04.009591 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-utilities\") pod \"redhat-marketplace-rhcsg\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:04 crc kubenswrapper[4799]: I1124 07:17:04.009753 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6jq6\" (UniqueName: \"kubernetes.io/projected/38de7098-5d03-4685-9ae3-0e96cd024915-kube-api-access-n6jq6\") pod \"redhat-marketplace-rhcsg\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:04 crc kubenswrapper[4799]: I1124 07:17:04.009988 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-catalog-content\") pod \"redhat-marketplace-rhcsg\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:04 crc kubenswrapper[4799]: I1124 07:17:04.010131 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-utilities\") pod \"redhat-marketplace-rhcsg\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:04 crc kubenswrapper[4799]: I1124 07:17:04.035520 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6jq6\" (UniqueName: \"kubernetes.io/projected/38de7098-5d03-4685-9ae3-0e96cd024915-kube-api-access-n6jq6\") pod \"redhat-marketplace-rhcsg\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:04 crc kubenswrapper[4799]: I1124 07:17:04.137944 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:04 crc kubenswrapper[4799]: I1124 07:17:04.568787 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rhcsg"] Nov 24 07:17:05 crc kubenswrapper[4799]: I1124 07:17:05.403297 4799 generic.go:334] "Generic (PLEG): container finished" podID="38de7098-5d03-4685-9ae3-0e96cd024915" containerID="9ae3c7833fc56cee16878898c5f53f87398b9705ef9ab90d8a108616fd603b7d" exitCode=0 Nov 24 07:17:05 crc kubenswrapper[4799]: I1124 07:17:05.403751 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhcsg" event={"ID":"38de7098-5d03-4685-9ae3-0e96cd024915","Type":"ContainerDied","Data":"9ae3c7833fc56cee16878898c5f53f87398b9705ef9ab90d8a108616fd603b7d"} Nov 24 07:17:05 crc kubenswrapper[4799]: I1124 07:17:05.403794 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhcsg" event={"ID":"38de7098-5d03-4685-9ae3-0e96cd024915","Type":"ContainerStarted","Data":"bbddbf2dfcf6450503df1c4c3768c04406153d8583287a0eead344a692952dba"} Nov 24 07:17:05 crc kubenswrapper[4799]: I1124 07:17:05.408522 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:17:06 crc kubenswrapper[4799]: I1124 07:17:06.415805 4799 generic.go:334] "Generic (PLEG): container finished" podID="38de7098-5d03-4685-9ae3-0e96cd024915" containerID="2bf627b15575b1990f340963946235b8ff06e6b0e834db82a03c2bab2cc0fed3" exitCode=0 Nov 24 07:17:06 crc kubenswrapper[4799]: I1124 07:17:06.415863 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhcsg" event={"ID":"38de7098-5d03-4685-9ae3-0e96cd024915","Type":"ContainerDied","Data":"2bf627b15575b1990f340963946235b8ff06e6b0e834db82a03c2bab2cc0fed3"} Nov 24 07:17:07 crc kubenswrapper[4799]: I1124 07:17:07.428432 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhcsg" event={"ID":"38de7098-5d03-4685-9ae3-0e96cd024915","Type":"ContainerStarted","Data":"6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092"} Nov 24 07:17:07 crc kubenswrapper[4799]: I1124 07:17:07.461566 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rhcsg" podStartSLOduration=3.025431976 podStartE2EDuration="4.461532522s" podCreationTimestamp="2025-11-24 07:17:03 +0000 UTC" firstStartedPulling="2025-11-24 07:17:05.408020894 +0000 UTC m=+1771.064003408" lastFinishedPulling="2025-11-24 07:17:06.84412145 +0000 UTC m=+1772.500103954" observedRunningTime="2025-11-24 07:17:07.45441427 +0000 UTC m=+1773.110396784" watchObservedRunningTime="2025-11-24 07:17:07.461532522 +0000 UTC m=+1773.117515036" Nov 24 07:17:12 crc kubenswrapper[4799]: I1124 07:17:12.628070 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:17:12 crc kubenswrapper[4799]: E1124 07:17:12.629217 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:17:14 crc kubenswrapper[4799]: I1124 07:17:14.139279 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:14 crc kubenswrapper[4799]: I1124 07:17:14.139337 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:14 crc kubenswrapper[4799]: I1124 07:17:14.210798 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:14 crc kubenswrapper[4799]: I1124 07:17:14.535745 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:14 crc kubenswrapper[4799]: I1124 07:17:14.919490 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rhcsg"] Nov 24 07:17:16 crc kubenswrapper[4799]: I1124 07:17:16.506336 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rhcsg" podUID="38de7098-5d03-4685-9ae3-0e96cd024915" containerName="registry-server" containerID="cri-o://6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092" gracePeriod=2 Nov 24 07:17:16 crc kubenswrapper[4799]: I1124 07:17:16.982743 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.152229 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-catalog-content\") pod \"38de7098-5d03-4685-9ae3-0e96cd024915\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.152301 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-utilities\") pod \"38de7098-5d03-4685-9ae3-0e96cd024915\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.152422 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6jq6\" (UniqueName: \"kubernetes.io/projected/38de7098-5d03-4685-9ae3-0e96cd024915-kube-api-access-n6jq6\") pod \"38de7098-5d03-4685-9ae3-0e96cd024915\" (UID: \"38de7098-5d03-4685-9ae3-0e96cd024915\") " Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.153489 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-utilities" (OuterVolumeSpecName: "utilities") pod "38de7098-5d03-4685-9ae3-0e96cd024915" (UID: "38de7098-5d03-4685-9ae3-0e96cd024915"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.158256 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38de7098-5d03-4685-9ae3-0e96cd024915-kube-api-access-n6jq6" (OuterVolumeSpecName: "kube-api-access-n6jq6") pod "38de7098-5d03-4685-9ae3-0e96cd024915" (UID: "38de7098-5d03-4685-9ae3-0e96cd024915"). InnerVolumeSpecName "kube-api-access-n6jq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.192154 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38de7098-5d03-4685-9ae3-0e96cd024915" (UID: "38de7098-5d03-4685-9ae3-0e96cd024915"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.254352 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6jq6\" (UniqueName: \"kubernetes.io/projected/38de7098-5d03-4685-9ae3-0e96cd024915-kube-api-access-n6jq6\") on node \"crc\" DevicePath \"\"" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.254399 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.254418 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38de7098-5d03-4685-9ae3-0e96cd024915-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.520950 4799 generic.go:334] "Generic (PLEG): container finished" podID="38de7098-5d03-4685-9ae3-0e96cd024915" containerID="6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092" exitCode=0 Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.521010 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rhcsg" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.521012 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhcsg" event={"ID":"38de7098-5d03-4685-9ae3-0e96cd024915","Type":"ContainerDied","Data":"6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092"} Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.521257 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhcsg" event={"ID":"38de7098-5d03-4685-9ae3-0e96cd024915","Type":"ContainerDied","Data":"bbddbf2dfcf6450503df1c4c3768c04406153d8583287a0eead344a692952dba"} Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.521308 4799 scope.go:117] "RemoveContainer" containerID="6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.552784 4799 scope.go:117] "RemoveContainer" containerID="2bf627b15575b1990f340963946235b8ff06e6b0e834db82a03c2bab2cc0fed3" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.570600 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rhcsg"] Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.587476 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rhcsg"] Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.602316 4799 scope.go:117] "RemoveContainer" containerID="9ae3c7833fc56cee16878898c5f53f87398b9705ef9ab90d8a108616fd603b7d" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.630021 4799 scope.go:117] "RemoveContainer" containerID="6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092" Nov 24 07:17:17 crc kubenswrapper[4799]: E1124 07:17:17.630611 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092\": container with ID starting with 6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092 not found: ID does not exist" containerID="6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.630651 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092"} err="failed to get container status \"6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092\": rpc error: code = NotFound desc = could not find container \"6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092\": container with ID starting with 6aa4c8edc15939667f1755138f560bca7ac896ad8a40ec82c3d2bb01a6fc5092 not found: ID does not exist" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.630680 4799 scope.go:117] "RemoveContainer" containerID="2bf627b15575b1990f340963946235b8ff06e6b0e834db82a03c2bab2cc0fed3" Nov 24 07:17:17 crc kubenswrapper[4799]: E1124 07:17:17.631342 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bf627b15575b1990f340963946235b8ff06e6b0e834db82a03c2bab2cc0fed3\": container with ID starting with 2bf627b15575b1990f340963946235b8ff06e6b0e834db82a03c2bab2cc0fed3 not found: ID does not exist" containerID="2bf627b15575b1990f340963946235b8ff06e6b0e834db82a03c2bab2cc0fed3" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.631405 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bf627b15575b1990f340963946235b8ff06e6b0e834db82a03c2bab2cc0fed3"} err="failed to get container status \"2bf627b15575b1990f340963946235b8ff06e6b0e834db82a03c2bab2cc0fed3\": rpc error: code = NotFound desc = could not find container \"2bf627b15575b1990f340963946235b8ff06e6b0e834db82a03c2bab2cc0fed3\": container with ID starting with 2bf627b15575b1990f340963946235b8ff06e6b0e834db82a03c2bab2cc0fed3 not found: ID does not exist" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.631450 4799 scope.go:117] "RemoveContainer" containerID="9ae3c7833fc56cee16878898c5f53f87398b9705ef9ab90d8a108616fd603b7d" Nov 24 07:17:17 crc kubenswrapper[4799]: E1124 07:17:17.631818 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ae3c7833fc56cee16878898c5f53f87398b9705ef9ab90d8a108616fd603b7d\": container with ID starting with 9ae3c7833fc56cee16878898c5f53f87398b9705ef9ab90d8a108616fd603b7d not found: ID does not exist" containerID="9ae3c7833fc56cee16878898c5f53f87398b9705ef9ab90d8a108616fd603b7d" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.631855 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ae3c7833fc56cee16878898c5f53f87398b9705ef9ab90d8a108616fd603b7d"} err="failed to get container status \"9ae3c7833fc56cee16878898c5f53f87398b9705ef9ab90d8a108616fd603b7d\": rpc error: code = NotFound desc = could not find container \"9ae3c7833fc56cee16878898c5f53f87398b9705ef9ab90d8a108616fd603b7d\": container with ID starting with 9ae3c7833fc56cee16878898c5f53f87398b9705ef9ab90d8a108616fd603b7d not found: ID does not exist" Nov 24 07:17:17 crc kubenswrapper[4799]: I1124 07:17:17.638320 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38de7098-5d03-4685-9ae3-0e96cd024915" path="/var/lib/kubelet/pods/38de7098-5d03-4685-9ae3-0e96cd024915/volumes" Nov 24 07:17:23 crc kubenswrapper[4799]: I1124 07:17:23.629313 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:17:23 crc kubenswrapper[4799]: E1124 07:17:23.630304 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:17:38 crc kubenswrapper[4799]: I1124 07:17:38.628585 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:17:38 crc kubenswrapper[4799]: E1124 07:17:38.629880 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:17:53 crc kubenswrapper[4799]: I1124 07:17:53.628714 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:17:53 crc kubenswrapper[4799]: I1124 07:17:53.896426 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"2d1ff395c59f441af8ed882cedd505590213ff88b8c90f275b0741bb8e5e9d22"} Nov 24 07:20:20 crc kubenswrapper[4799]: I1124 07:20:20.401755 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:20:20 crc kubenswrapper[4799]: I1124 07:20:20.402569 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:20:50 crc kubenswrapper[4799]: I1124 07:20:50.400921 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:20:50 crc kubenswrapper[4799]: I1124 07:20:50.401378 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:21:20 crc kubenswrapper[4799]: I1124 07:21:20.401194 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:21:20 crc kubenswrapper[4799]: I1124 07:21:20.403071 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:21:20 crc kubenswrapper[4799]: I1124 07:21:20.403159 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:21:20 crc kubenswrapper[4799]: I1124 07:21:20.404079 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2d1ff395c59f441af8ed882cedd505590213ff88b8c90f275b0741bb8e5e9d22"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:21:20 crc kubenswrapper[4799]: I1124 07:21:20.404180 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://2d1ff395c59f441af8ed882cedd505590213ff88b8c90f275b0741bb8e5e9d22" gracePeriod=600 Nov 24 07:21:20 crc kubenswrapper[4799]: I1124 07:21:20.867835 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="2d1ff395c59f441af8ed882cedd505590213ff88b8c90f275b0741bb8e5e9d22" exitCode=0 Nov 24 07:21:20 crc kubenswrapper[4799]: I1124 07:21:20.867913 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"2d1ff395c59f441af8ed882cedd505590213ff88b8c90f275b0741bb8e5e9d22"} Nov 24 07:21:20 crc kubenswrapper[4799]: I1124 07:21:20.867957 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce"} Nov 24 07:21:20 crc kubenswrapper[4799]: I1124 07:21:20.867979 4799 scope.go:117] "RemoveContainer" containerID="f6d308b393e46984fb017cb515be934b152cd9e0187a1b50dd33187f669674e6" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.474189 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-982w5"] Nov 24 07:22:09 crc kubenswrapper[4799]: E1124 07:22:09.475403 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38de7098-5d03-4685-9ae3-0e96cd024915" containerName="extract-utilities" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.475432 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="38de7098-5d03-4685-9ae3-0e96cd024915" containerName="extract-utilities" Nov 24 07:22:09 crc kubenswrapper[4799]: E1124 07:22:09.475456 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38de7098-5d03-4685-9ae3-0e96cd024915" containerName="extract-content" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.475468 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="38de7098-5d03-4685-9ae3-0e96cd024915" containerName="extract-content" Nov 24 07:22:09 crc kubenswrapper[4799]: E1124 07:22:09.475495 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38de7098-5d03-4685-9ae3-0e96cd024915" containerName="registry-server" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.475510 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="38de7098-5d03-4685-9ae3-0e96cd024915" containerName="registry-server" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.475754 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="38de7098-5d03-4685-9ae3-0e96cd024915" containerName="registry-server" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.480314 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.518402 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-982w5"] Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.576196 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvxx9\" (UniqueName: \"kubernetes.io/projected/ef582943-e853-42ef-a86c-f655035e4ca7-kube-api-access-xvxx9\") pod \"community-operators-982w5\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.576553 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-catalog-content\") pod \"community-operators-982w5\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.576687 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-utilities\") pod \"community-operators-982w5\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.678045 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvxx9\" (UniqueName: \"kubernetes.io/projected/ef582943-e853-42ef-a86c-f655035e4ca7-kube-api-access-xvxx9\") pod \"community-operators-982w5\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.678110 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-catalog-content\") pod \"community-operators-982w5\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.678154 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-utilities\") pod \"community-operators-982w5\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.678764 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-utilities\") pod \"community-operators-982w5\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.678774 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-catalog-content\") pod \"community-operators-982w5\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.709535 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvxx9\" (UniqueName: \"kubernetes.io/projected/ef582943-e853-42ef-a86c-f655035e4ca7-kube-api-access-xvxx9\") pod \"community-operators-982w5\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:09 crc kubenswrapper[4799]: I1124 07:22:09.834725 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:10 crc kubenswrapper[4799]: I1124 07:22:10.125181 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-982w5"] Nov 24 07:22:10 crc kubenswrapper[4799]: I1124 07:22:10.350334 4799 generic.go:334] "Generic (PLEG): container finished" podID="ef582943-e853-42ef-a86c-f655035e4ca7" containerID="6654ebf79d3243daa7909367a28ea77285c0ade9e6f82c6227fd111fb9de87e0" exitCode=0 Nov 24 07:22:10 crc kubenswrapper[4799]: I1124 07:22:10.350437 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-982w5" event={"ID":"ef582943-e853-42ef-a86c-f655035e4ca7","Type":"ContainerDied","Data":"6654ebf79d3243daa7909367a28ea77285c0ade9e6f82c6227fd111fb9de87e0"} Nov 24 07:22:10 crc kubenswrapper[4799]: I1124 07:22:10.350717 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-982w5" event={"ID":"ef582943-e853-42ef-a86c-f655035e4ca7","Type":"ContainerStarted","Data":"be7f87b2cbf0674cc5abec79bb4bacc4b7945693fee8e8fce96f85f0e6c9173a"} Nov 24 07:22:10 crc kubenswrapper[4799]: I1124 07:22:10.353176 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:22:11 crc kubenswrapper[4799]: I1124 07:22:11.370761 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-982w5" event={"ID":"ef582943-e853-42ef-a86c-f655035e4ca7","Type":"ContainerStarted","Data":"94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9"} Nov 24 07:22:12 crc kubenswrapper[4799]: I1124 07:22:12.386472 4799 generic.go:334] "Generic (PLEG): container finished" podID="ef582943-e853-42ef-a86c-f655035e4ca7" containerID="94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9" exitCode=0 Nov 24 07:22:12 crc kubenswrapper[4799]: I1124 07:22:12.386535 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-982w5" event={"ID":"ef582943-e853-42ef-a86c-f655035e4ca7","Type":"ContainerDied","Data":"94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9"} Nov 24 07:22:13 crc kubenswrapper[4799]: I1124 07:22:13.400991 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-982w5" event={"ID":"ef582943-e853-42ef-a86c-f655035e4ca7","Type":"ContainerStarted","Data":"a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174"} Nov 24 07:22:13 crc kubenswrapper[4799]: I1124 07:22:13.440912 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-982w5" podStartSLOduration=1.958583095 podStartE2EDuration="4.44083438s" podCreationTimestamp="2025-11-24 07:22:09 +0000 UTC" firstStartedPulling="2025-11-24 07:22:10.35279792 +0000 UTC m=+2076.008780414" lastFinishedPulling="2025-11-24 07:22:12.835049175 +0000 UTC m=+2078.491031699" observedRunningTime="2025-11-24 07:22:13.431884414 +0000 UTC m=+2079.087866958" watchObservedRunningTime="2025-11-24 07:22:13.44083438 +0000 UTC m=+2079.096816894" Nov 24 07:22:19 crc kubenswrapper[4799]: I1124 07:22:19.836909 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:19 crc kubenswrapper[4799]: I1124 07:22:19.837691 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:19 crc kubenswrapper[4799]: I1124 07:22:19.902373 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:20 crc kubenswrapper[4799]: I1124 07:22:20.540805 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:20 crc kubenswrapper[4799]: I1124 07:22:20.610454 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-982w5"] Nov 24 07:22:22 crc kubenswrapper[4799]: I1124 07:22:22.486750 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-982w5" podUID="ef582943-e853-42ef-a86c-f655035e4ca7" containerName="registry-server" containerID="cri-o://a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174" gracePeriod=2 Nov 24 07:22:22 crc kubenswrapper[4799]: I1124 07:22:22.971191 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.100138 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvxx9\" (UniqueName: \"kubernetes.io/projected/ef582943-e853-42ef-a86c-f655035e4ca7-kube-api-access-xvxx9\") pod \"ef582943-e853-42ef-a86c-f655035e4ca7\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.100429 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-catalog-content\") pod \"ef582943-e853-42ef-a86c-f655035e4ca7\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.100595 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-utilities\") pod \"ef582943-e853-42ef-a86c-f655035e4ca7\" (UID: \"ef582943-e853-42ef-a86c-f655035e4ca7\") " Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.102080 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-utilities" (OuterVolumeSpecName: "utilities") pod "ef582943-e853-42ef-a86c-f655035e4ca7" (UID: "ef582943-e853-42ef-a86c-f655035e4ca7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.109503 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef582943-e853-42ef-a86c-f655035e4ca7-kube-api-access-xvxx9" (OuterVolumeSpecName: "kube-api-access-xvxx9") pod "ef582943-e853-42ef-a86c-f655035e4ca7" (UID: "ef582943-e853-42ef-a86c-f655035e4ca7"). InnerVolumeSpecName "kube-api-access-xvxx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.151725 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef582943-e853-42ef-a86c-f655035e4ca7" (UID: "ef582943-e853-42ef-a86c-f655035e4ca7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.202642 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvxx9\" (UniqueName: \"kubernetes.io/projected/ef582943-e853-42ef-a86c-f655035e4ca7-kube-api-access-xvxx9\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.202678 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.202690 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef582943-e853-42ef-a86c-f655035e4ca7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.504257 4799 generic.go:334] "Generic (PLEG): container finished" podID="ef582943-e853-42ef-a86c-f655035e4ca7" containerID="a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174" exitCode=0 Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.504341 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-982w5" event={"ID":"ef582943-e853-42ef-a86c-f655035e4ca7","Type":"ContainerDied","Data":"a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174"} Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.504393 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-982w5" event={"ID":"ef582943-e853-42ef-a86c-f655035e4ca7","Type":"ContainerDied","Data":"be7f87b2cbf0674cc5abec79bb4bacc4b7945693fee8e8fce96f85f0e6c9173a"} Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.504350 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-982w5" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.504429 4799 scope.go:117] "RemoveContainer" containerID="a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.554829 4799 scope.go:117] "RemoveContainer" containerID="94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.568146 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-982w5"] Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.579528 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-982w5"] Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.586930 4799 scope.go:117] "RemoveContainer" containerID="6654ebf79d3243daa7909367a28ea77285c0ade9e6f82c6227fd111fb9de87e0" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.621389 4799 scope.go:117] "RemoveContainer" containerID="a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174" Nov 24 07:22:23 crc kubenswrapper[4799]: E1124 07:22:23.622186 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174\": container with ID starting with a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174 not found: ID does not exist" containerID="a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.622268 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174"} err="failed to get container status \"a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174\": rpc error: code = NotFound desc = could not find container \"a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174\": container with ID starting with a6f7fb562bc908f9295fba827d92c1ebd66a2122d20187227c7297be42544174 not found: ID does not exist" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.622315 4799 scope.go:117] "RemoveContainer" containerID="94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9" Nov 24 07:22:23 crc kubenswrapper[4799]: E1124 07:22:23.622917 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9\": container with ID starting with 94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9 not found: ID does not exist" containerID="94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.622986 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9"} err="failed to get container status \"94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9\": rpc error: code = NotFound desc = could not find container \"94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9\": container with ID starting with 94cbac4118adf3fa37711854f1191649cf224edc88604ff6fcfc095ff8f043b9 not found: ID does not exist" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.623018 4799 scope.go:117] "RemoveContainer" containerID="6654ebf79d3243daa7909367a28ea77285c0ade9e6f82c6227fd111fb9de87e0" Nov 24 07:22:23 crc kubenswrapper[4799]: E1124 07:22:23.623762 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6654ebf79d3243daa7909367a28ea77285c0ade9e6f82c6227fd111fb9de87e0\": container with ID starting with 6654ebf79d3243daa7909367a28ea77285c0ade9e6f82c6227fd111fb9de87e0 not found: ID does not exist" containerID="6654ebf79d3243daa7909367a28ea77285c0ade9e6f82c6227fd111fb9de87e0" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.623809 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6654ebf79d3243daa7909367a28ea77285c0ade9e6f82c6227fd111fb9de87e0"} err="failed to get container status \"6654ebf79d3243daa7909367a28ea77285c0ade9e6f82c6227fd111fb9de87e0\": rpc error: code = NotFound desc = could not find container \"6654ebf79d3243daa7909367a28ea77285c0ade9e6f82c6227fd111fb9de87e0\": container with ID starting with 6654ebf79d3243daa7909367a28ea77285c0ade9e6f82c6227fd111fb9de87e0 not found: ID does not exist" Nov 24 07:22:23 crc kubenswrapper[4799]: I1124 07:22:23.643798 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef582943-e853-42ef-a86c-f655035e4ca7" path="/var/lib/kubelet/pods/ef582943-e853-42ef-a86c-f655035e4ca7/volumes" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.600827 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2l2kb"] Nov 24 07:22:42 crc kubenswrapper[4799]: E1124 07:22:42.601669 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef582943-e853-42ef-a86c-f655035e4ca7" containerName="registry-server" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.601684 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef582943-e853-42ef-a86c-f655035e4ca7" containerName="registry-server" Nov 24 07:22:42 crc kubenswrapper[4799]: E1124 07:22:42.601705 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef582943-e853-42ef-a86c-f655035e4ca7" containerName="extract-content" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.601713 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef582943-e853-42ef-a86c-f655035e4ca7" containerName="extract-content" Nov 24 07:22:42 crc kubenswrapper[4799]: E1124 07:22:42.601735 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef582943-e853-42ef-a86c-f655035e4ca7" containerName="extract-utilities" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.601744 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef582943-e853-42ef-a86c-f655035e4ca7" containerName="extract-utilities" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.601947 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef582943-e853-42ef-a86c-f655035e4ca7" containerName="registry-server" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.603121 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.620586 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2l2kb"] Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.752878 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwtnc\" (UniqueName: \"kubernetes.io/projected/918414fc-235b-446a-b5f9-b1ee6530d040-kube-api-access-pwtnc\") pod \"certified-operators-2l2kb\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.752951 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-catalog-content\") pod \"certified-operators-2l2kb\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.752986 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-utilities\") pod \"certified-operators-2l2kb\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.855641 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwtnc\" (UniqueName: \"kubernetes.io/projected/918414fc-235b-446a-b5f9-b1ee6530d040-kube-api-access-pwtnc\") pod \"certified-operators-2l2kb\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.855744 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-catalog-content\") pod \"certified-operators-2l2kb\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.855799 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-utilities\") pod \"certified-operators-2l2kb\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.856640 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-catalog-content\") pod \"certified-operators-2l2kb\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.856751 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-utilities\") pod \"certified-operators-2l2kb\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.893787 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwtnc\" (UniqueName: \"kubernetes.io/projected/918414fc-235b-446a-b5f9-b1ee6530d040-kube-api-access-pwtnc\") pod \"certified-operators-2l2kb\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:42 crc kubenswrapper[4799]: I1124 07:22:42.939196 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:43 crc kubenswrapper[4799]: I1124 07:22:43.243291 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2l2kb"] Nov 24 07:22:43 crc kubenswrapper[4799]: W1124 07:22:43.256989 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod918414fc_235b_446a_b5f9_b1ee6530d040.slice/crio-e4e07ae522bf8c9f0d092aca7121ddbd3915b8419a095e858a829dca97f7fac5 WatchSource:0}: Error finding container e4e07ae522bf8c9f0d092aca7121ddbd3915b8419a095e858a829dca97f7fac5: Status 404 returned error can't find the container with id e4e07ae522bf8c9f0d092aca7121ddbd3915b8419a095e858a829dca97f7fac5 Nov 24 07:22:43 crc kubenswrapper[4799]: I1124 07:22:43.713725 4799 generic.go:334] "Generic (PLEG): container finished" podID="918414fc-235b-446a-b5f9-b1ee6530d040" containerID="3fdc3ea1b9caf3c095d05001522430e8bf0bc6819335c0fd570dd8c8b7b581ee" exitCode=0 Nov 24 07:22:43 crc kubenswrapper[4799]: I1124 07:22:43.713841 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2kb" event={"ID":"918414fc-235b-446a-b5f9-b1ee6530d040","Type":"ContainerDied","Data":"3fdc3ea1b9caf3c095d05001522430e8bf0bc6819335c0fd570dd8c8b7b581ee"} Nov 24 07:22:43 crc kubenswrapper[4799]: I1124 07:22:43.714543 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2kb" event={"ID":"918414fc-235b-446a-b5f9-b1ee6530d040","Type":"ContainerStarted","Data":"e4e07ae522bf8c9f0d092aca7121ddbd3915b8419a095e858a829dca97f7fac5"} Nov 24 07:22:44 crc kubenswrapper[4799]: I1124 07:22:44.726785 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2kb" event={"ID":"918414fc-235b-446a-b5f9-b1ee6530d040","Type":"ContainerStarted","Data":"7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879"} Nov 24 07:22:45 crc kubenswrapper[4799]: I1124 07:22:45.735378 4799 generic.go:334] "Generic (PLEG): container finished" podID="918414fc-235b-446a-b5f9-b1ee6530d040" containerID="7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879" exitCode=0 Nov 24 07:22:45 crc kubenswrapper[4799]: I1124 07:22:45.735499 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2kb" event={"ID":"918414fc-235b-446a-b5f9-b1ee6530d040","Type":"ContainerDied","Data":"7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879"} Nov 24 07:22:46 crc kubenswrapper[4799]: I1124 07:22:46.748747 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2kb" event={"ID":"918414fc-235b-446a-b5f9-b1ee6530d040","Type":"ContainerStarted","Data":"25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12"} Nov 24 07:22:46 crc kubenswrapper[4799]: I1124 07:22:46.767611 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2l2kb" podStartSLOduration=2.33350673 podStartE2EDuration="4.767585415s" podCreationTimestamp="2025-11-24 07:22:42 +0000 UTC" firstStartedPulling="2025-11-24 07:22:43.72582339 +0000 UTC m=+2109.381805904" lastFinishedPulling="2025-11-24 07:22:46.159902045 +0000 UTC m=+2111.815884589" observedRunningTime="2025-11-24 07:22:46.766133333 +0000 UTC m=+2112.422115847" watchObservedRunningTime="2025-11-24 07:22:46.767585415 +0000 UTC m=+2112.423567919" Nov 24 07:22:52 crc kubenswrapper[4799]: I1124 07:22:52.939470 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:52 crc kubenswrapper[4799]: I1124 07:22:52.939934 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:53 crc kubenswrapper[4799]: I1124 07:22:53.017308 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:53 crc kubenswrapper[4799]: I1124 07:22:53.872708 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:53 crc kubenswrapper[4799]: I1124 07:22:53.926188 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2l2kb"] Nov 24 07:22:55 crc kubenswrapper[4799]: I1124 07:22:55.832691 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2l2kb" podUID="918414fc-235b-446a-b5f9-b1ee6530d040" containerName="registry-server" containerID="cri-o://25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12" gracePeriod=2 Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.276221 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.375558 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-catalog-content\") pod \"918414fc-235b-446a-b5f9-b1ee6530d040\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.375649 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-utilities\") pod \"918414fc-235b-446a-b5f9-b1ee6530d040\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.375715 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwtnc\" (UniqueName: \"kubernetes.io/projected/918414fc-235b-446a-b5f9-b1ee6530d040-kube-api-access-pwtnc\") pod \"918414fc-235b-446a-b5f9-b1ee6530d040\" (UID: \"918414fc-235b-446a-b5f9-b1ee6530d040\") " Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.377053 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-utilities" (OuterVolumeSpecName: "utilities") pod "918414fc-235b-446a-b5f9-b1ee6530d040" (UID: "918414fc-235b-446a-b5f9-b1ee6530d040"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.384029 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/918414fc-235b-446a-b5f9-b1ee6530d040-kube-api-access-pwtnc" (OuterVolumeSpecName: "kube-api-access-pwtnc") pod "918414fc-235b-446a-b5f9-b1ee6530d040" (UID: "918414fc-235b-446a-b5f9-b1ee6530d040"). InnerVolumeSpecName "kube-api-access-pwtnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.473744 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "918414fc-235b-446a-b5f9-b1ee6530d040" (UID: "918414fc-235b-446a-b5f9-b1ee6530d040"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.478075 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.478223 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/918414fc-235b-446a-b5f9-b1ee6530d040-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.478301 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwtnc\" (UniqueName: \"kubernetes.io/projected/918414fc-235b-446a-b5f9-b1ee6530d040-kube-api-access-pwtnc\") on node \"crc\" DevicePath \"\"" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.846987 4799 generic.go:334] "Generic (PLEG): container finished" podID="918414fc-235b-446a-b5f9-b1ee6530d040" containerID="25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12" exitCode=0 Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.847059 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2kb" event={"ID":"918414fc-235b-446a-b5f9-b1ee6530d040","Type":"ContainerDied","Data":"25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12"} Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.847103 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2kb" event={"ID":"918414fc-235b-446a-b5f9-b1ee6530d040","Type":"ContainerDied","Data":"e4e07ae522bf8c9f0d092aca7121ddbd3915b8419a095e858a829dca97f7fac5"} Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.847105 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l2kb" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.847135 4799 scope.go:117] "RemoveContainer" containerID="25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.883597 4799 scope.go:117] "RemoveContainer" containerID="7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.923712 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2l2kb"] Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.928453 4799 scope.go:117] "RemoveContainer" containerID="3fdc3ea1b9caf3c095d05001522430e8bf0bc6819335c0fd570dd8c8b7b581ee" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.931800 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2l2kb"] Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.956787 4799 scope.go:117] "RemoveContainer" containerID="25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12" Nov 24 07:22:56 crc kubenswrapper[4799]: E1124 07:22:56.957290 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12\": container with ID starting with 25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12 not found: ID does not exist" containerID="25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.957346 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12"} err="failed to get container status \"25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12\": rpc error: code = NotFound desc = could not find container \"25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12\": container with ID starting with 25d3567b9712f33b238ddbb958a7c6137a7a123a74ff15880277a8c86d33bd12 not found: ID does not exist" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.957386 4799 scope.go:117] "RemoveContainer" containerID="7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879" Nov 24 07:22:56 crc kubenswrapper[4799]: E1124 07:22:56.957824 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879\": container with ID starting with 7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879 not found: ID does not exist" containerID="7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.957889 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879"} err="failed to get container status \"7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879\": rpc error: code = NotFound desc = could not find container \"7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879\": container with ID starting with 7e09abccc9f32ea4b916ab8e9558f067e6cc3f807852dc26f5c3badb549de879 not found: ID does not exist" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.957919 4799 scope.go:117] "RemoveContainer" containerID="3fdc3ea1b9caf3c095d05001522430e8bf0bc6819335c0fd570dd8c8b7b581ee" Nov 24 07:22:56 crc kubenswrapper[4799]: E1124 07:22:56.958489 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fdc3ea1b9caf3c095d05001522430e8bf0bc6819335c0fd570dd8c8b7b581ee\": container with ID starting with 3fdc3ea1b9caf3c095d05001522430e8bf0bc6819335c0fd570dd8c8b7b581ee not found: ID does not exist" containerID="3fdc3ea1b9caf3c095d05001522430e8bf0bc6819335c0fd570dd8c8b7b581ee" Nov 24 07:22:56 crc kubenswrapper[4799]: I1124 07:22:56.958542 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fdc3ea1b9caf3c095d05001522430e8bf0bc6819335c0fd570dd8c8b7b581ee"} err="failed to get container status \"3fdc3ea1b9caf3c095d05001522430e8bf0bc6819335c0fd570dd8c8b7b581ee\": rpc error: code = NotFound desc = could not find container \"3fdc3ea1b9caf3c095d05001522430e8bf0bc6819335c0fd570dd8c8b7b581ee\": container with ID starting with 3fdc3ea1b9caf3c095d05001522430e8bf0bc6819335c0fd570dd8c8b7b581ee not found: ID does not exist" Nov 24 07:22:57 crc kubenswrapper[4799]: I1124 07:22:57.646148 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="918414fc-235b-446a-b5f9-b1ee6530d040" path="/var/lib/kubelet/pods/918414fc-235b-446a-b5f9-b1ee6530d040/volumes" Nov 24 07:23:20 crc kubenswrapper[4799]: I1124 07:23:20.400324 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:23:20 crc kubenswrapper[4799]: I1124 07:23:20.401201 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:23:50 crc kubenswrapper[4799]: I1124 07:23:50.401550 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:23:50 crc kubenswrapper[4799]: I1124 07:23:50.403135 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:24:20 crc kubenswrapper[4799]: I1124 07:24:20.401688 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:24:20 crc kubenswrapper[4799]: I1124 07:24:20.403108 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:24:20 crc kubenswrapper[4799]: I1124 07:24:20.403238 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:24:20 crc kubenswrapper[4799]: I1124 07:24:20.404590 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:24:20 crc kubenswrapper[4799]: I1124 07:24:20.404745 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" gracePeriod=600 Nov 24 07:24:20 crc kubenswrapper[4799]: E1124 07:24:20.534831 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:24:20 crc kubenswrapper[4799]: I1124 07:24:20.725379 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" exitCode=0 Nov 24 07:24:20 crc kubenswrapper[4799]: I1124 07:24:20.725426 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce"} Nov 24 07:24:20 crc kubenswrapper[4799]: I1124 07:24:20.725468 4799 scope.go:117] "RemoveContainer" containerID="2d1ff395c59f441af8ed882cedd505590213ff88b8c90f275b0741bb8e5e9d22" Nov 24 07:24:20 crc kubenswrapper[4799]: I1124 07:24:20.725951 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:24:20 crc kubenswrapper[4799]: E1124 07:24:20.726200 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.589108 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d6c4s"] Nov 24 07:24:21 crc kubenswrapper[4799]: E1124 07:24:21.589614 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="918414fc-235b-446a-b5f9-b1ee6530d040" containerName="extract-content" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.589640 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="918414fc-235b-446a-b5f9-b1ee6530d040" containerName="extract-content" Nov 24 07:24:21 crc kubenswrapper[4799]: E1124 07:24:21.589696 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="918414fc-235b-446a-b5f9-b1ee6530d040" containerName="registry-server" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.589712 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="918414fc-235b-446a-b5f9-b1ee6530d040" containerName="registry-server" Nov 24 07:24:21 crc kubenswrapper[4799]: E1124 07:24:21.589741 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="918414fc-235b-446a-b5f9-b1ee6530d040" containerName="extract-utilities" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.589755 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="918414fc-235b-446a-b5f9-b1ee6530d040" containerName="extract-utilities" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.591895 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="918414fc-235b-446a-b5f9-b1ee6530d040" containerName="registry-server" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.593955 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.604217 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d6c4s"] Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.700499 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-utilities\") pod \"redhat-operators-d6c4s\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.700584 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldrm9\" (UniqueName: \"kubernetes.io/projected/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-kube-api-access-ldrm9\") pod \"redhat-operators-d6c4s\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.700613 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-catalog-content\") pod \"redhat-operators-d6c4s\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.802153 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-utilities\") pod \"redhat-operators-d6c4s\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.802293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldrm9\" (UniqueName: \"kubernetes.io/projected/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-kube-api-access-ldrm9\") pod \"redhat-operators-d6c4s\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.802340 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-catalog-content\") pod \"redhat-operators-d6c4s\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.802791 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-utilities\") pod \"redhat-operators-d6c4s\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.803185 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-catalog-content\") pod \"redhat-operators-d6c4s\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.819692 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldrm9\" (UniqueName: \"kubernetes.io/projected/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-kube-api-access-ldrm9\") pod \"redhat-operators-d6c4s\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:21 crc kubenswrapper[4799]: I1124 07:24:21.969897 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:22 crc kubenswrapper[4799]: I1124 07:24:22.397911 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d6c4s"] Nov 24 07:24:22 crc kubenswrapper[4799]: I1124 07:24:22.742381 4799 generic.go:334] "Generic (PLEG): container finished" podID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" containerID="f3d4ec762043800af3d85f5599908dfc5a4437a391eb646b7db8306031bf90cd" exitCode=0 Nov 24 07:24:22 crc kubenswrapper[4799]: I1124 07:24:22.742428 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6c4s" event={"ID":"67d82441-ef4d-4711-ad7c-cd7aa11d9c39","Type":"ContainerDied","Data":"f3d4ec762043800af3d85f5599908dfc5a4437a391eb646b7db8306031bf90cd"} Nov 24 07:24:22 crc kubenswrapper[4799]: I1124 07:24:22.742631 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6c4s" event={"ID":"67d82441-ef4d-4711-ad7c-cd7aa11d9c39","Type":"ContainerStarted","Data":"dc554318068f2e337ee3704f993c63e9fde3d134c7de788476669e5ae56dd5b1"} Nov 24 07:24:23 crc kubenswrapper[4799]: I1124 07:24:23.753538 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6c4s" event={"ID":"67d82441-ef4d-4711-ad7c-cd7aa11d9c39","Type":"ContainerStarted","Data":"4e02af5c9f551488035e9dc955fdb57d2f2e5661a5f57f78425b59d6921c963a"} Nov 24 07:24:24 crc kubenswrapper[4799]: I1124 07:24:24.761149 4799 generic.go:334] "Generic (PLEG): container finished" podID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" containerID="4e02af5c9f551488035e9dc955fdb57d2f2e5661a5f57f78425b59d6921c963a" exitCode=0 Nov 24 07:24:24 crc kubenswrapper[4799]: I1124 07:24:24.761231 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6c4s" event={"ID":"67d82441-ef4d-4711-ad7c-cd7aa11d9c39","Type":"ContainerDied","Data":"4e02af5c9f551488035e9dc955fdb57d2f2e5661a5f57f78425b59d6921c963a"} Nov 24 07:24:25 crc kubenswrapper[4799]: I1124 07:24:25.769931 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6c4s" event={"ID":"67d82441-ef4d-4711-ad7c-cd7aa11d9c39","Type":"ContainerStarted","Data":"ce2bafea0da986c506a5e4bb01f601c99c635b95c99ab6e14b4c62e8d2007409"} Nov 24 07:24:25 crc kubenswrapper[4799]: I1124 07:24:25.791041 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d6c4s" podStartSLOduration=2.387373711 podStartE2EDuration="4.791025507s" podCreationTimestamp="2025-11-24 07:24:21 +0000 UTC" firstStartedPulling="2025-11-24 07:24:22.743860597 +0000 UTC m=+2208.399843071" lastFinishedPulling="2025-11-24 07:24:25.147512393 +0000 UTC m=+2210.803494867" observedRunningTime="2025-11-24 07:24:25.787809162 +0000 UTC m=+2211.443791636" watchObservedRunningTime="2025-11-24 07:24:25.791025507 +0000 UTC m=+2211.447007981" Nov 24 07:24:31 crc kubenswrapper[4799]: I1124 07:24:31.970567 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:31 crc kubenswrapper[4799]: I1124 07:24:31.971275 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:32 crc kubenswrapper[4799]: I1124 07:24:32.031717 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:32 crc kubenswrapper[4799]: I1124 07:24:32.867262 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:32 crc kubenswrapper[4799]: I1124 07:24:32.906787 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d6c4s"] Nov 24 07:24:33 crc kubenswrapper[4799]: I1124 07:24:33.628571 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:24:33 crc kubenswrapper[4799]: E1124 07:24:33.629217 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:24:34 crc kubenswrapper[4799]: I1124 07:24:34.839434 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d6c4s" podUID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" containerName="registry-server" containerID="cri-o://ce2bafea0da986c506a5e4bb01f601c99c635b95c99ab6e14b4c62e8d2007409" gracePeriod=2 Nov 24 07:24:35 crc kubenswrapper[4799]: I1124 07:24:35.852323 4799 generic.go:334] "Generic (PLEG): container finished" podID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" containerID="ce2bafea0da986c506a5e4bb01f601c99c635b95c99ab6e14b4c62e8d2007409" exitCode=0 Nov 24 07:24:35 crc kubenswrapper[4799]: I1124 07:24:35.852382 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6c4s" event={"ID":"67d82441-ef4d-4711-ad7c-cd7aa11d9c39","Type":"ContainerDied","Data":"ce2bafea0da986c506a5e4bb01f601c99c635b95c99ab6e14b4c62e8d2007409"} Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.207702 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.318109 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-catalog-content\") pod \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.318159 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-utilities\") pod \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.318211 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldrm9\" (UniqueName: \"kubernetes.io/projected/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-kube-api-access-ldrm9\") pod \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\" (UID: \"67d82441-ef4d-4711-ad7c-cd7aa11d9c39\") " Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.318820 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-utilities" (OuterVolumeSpecName: "utilities") pod "67d82441-ef4d-4711-ad7c-cd7aa11d9c39" (UID: "67d82441-ef4d-4711-ad7c-cd7aa11d9c39"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.326681 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-kube-api-access-ldrm9" (OuterVolumeSpecName: "kube-api-access-ldrm9") pod "67d82441-ef4d-4711-ad7c-cd7aa11d9c39" (UID: "67d82441-ef4d-4711-ad7c-cd7aa11d9c39"). InnerVolumeSpecName "kube-api-access-ldrm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.404988 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67d82441-ef4d-4711-ad7c-cd7aa11d9c39" (UID: "67d82441-ef4d-4711-ad7c-cd7aa11d9c39"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.420720 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.420747 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.420759 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldrm9\" (UniqueName: \"kubernetes.io/projected/67d82441-ef4d-4711-ad7c-cd7aa11d9c39-kube-api-access-ldrm9\") on node \"crc\" DevicePath \"\"" Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.865944 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6c4s" event={"ID":"67d82441-ef4d-4711-ad7c-cd7aa11d9c39","Type":"ContainerDied","Data":"dc554318068f2e337ee3704f993c63e9fde3d134c7de788476669e5ae56dd5b1"} Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.866027 4799 scope.go:117] "RemoveContainer" containerID="ce2bafea0da986c506a5e4bb01f601c99c635b95c99ab6e14b4c62e8d2007409" Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.866086 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6c4s" Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.898324 4799 scope.go:117] "RemoveContainer" containerID="4e02af5c9f551488035e9dc955fdb57d2f2e5661a5f57f78425b59d6921c963a" Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.917304 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d6c4s"] Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.926566 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d6c4s"] Nov 24 07:24:36 crc kubenswrapper[4799]: I1124 07:24:36.932515 4799 scope.go:117] "RemoveContainer" containerID="f3d4ec762043800af3d85f5599908dfc5a4437a391eb646b7db8306031bf90cd" Nov 24 07:24:37 crc kubenswrapper[4799]: I1124 07:24:37.635797 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" path="/var/lib/kubelet/pods/67d82441-ef4d-4711-ad7c-cd7aa11d9c39/volumes" Nov 24 07:24:47 crc kubenswrapper[4799]: I1124 07:24:47.628554 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:24:47 crc kubenswrapper[4799]: E1124 07:24:47.629747 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:25:00 crc kubenswrapper[4799]: I1124 07:25:00.627995 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:25:00 crc kubenswrapper[4799]: E1124 07:25:00.628735 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:25:13 crc kubenswrapper[4799]: I1124 07:25:13.629365 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:25:13 crc kubenswrapper[4799]: E1124 07:25:13.630493 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:25:24 crc kubenswrapper[4799]: I1124 07:25:24.628283 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:25:24 crc kubenswrapper[4799]: E1124 07:25:24.629424 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:25:38 crc kubenswrapper[4799]: I1124 07:25:38.628750 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:25:38 crc kubenswrapper[4799]: E1124 07:25:38.630373 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:25:52 crc kubenswrapper[4799]: I1124 07:25:52.629281 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:25:52 crc kubenswrapper[4799]: E1124 07:25:52.630247 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:26:06 crc kubenswrapper[4799]: I1124 07:26:06.628677 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:26:06 crc kubenswrapper[4799]: E1124 07:26:06.629802 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:26:17 crc kubenswrapper[4799]: I1124 07:26:17.628001 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:26:17 crc kubenswrapper[4799]: E1124 07:26:17.628528 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:26:32 crc kubenswrapper[4799]: I1124 07:26:32.627900 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:26:32 crc kubenswrapper[4799]: E1124 07:26:32.630041 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:26:47 crc kubenswrapper[4799]: I1124 07:26:47.629377 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:26:47 crc kubenswrapper[4799]: E1124 07:26:47.630034 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:26:59 crc kubenswrapper[4799]: I1124 07:26:59.628101 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:26:59 crc kubenswrapper[4799]: E1124 07:26:59.629235 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:27:10 crc kubenswrapper[4799]: I1124 07:27:10.629590 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:27:10 crc kubenswrapper[4799]: E1124 07:27:10.630617 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:27:23 crc kubenswrapper[4799]: I1124 07:27:23.629262 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:27:23 crc kubenswrapper[4799]: E1124 07:27:23.630558 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:27:35 crc kubenswrapper[4799]: I1124 07:27:35.632836 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:27:35 crc kubenswrapper[4799]: E1124 07:27:35.634212 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:27:48 crc kubenswrapper[4799]: I1124 07:27:48.628703 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:27:48 crc kubenswrapper[4799]: E1124 07:27:48.629567 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:28:03 crc kubenswrapper[4799]: I1124 07:28:03.628860 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:28:03 crc kubenswrapper[4799]: E1124 07:28:03.629752 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:28:16 crc kubenswrapper[4799]: I1124 07:28:16.628471 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:28:16 crc kubenswrapper[4799]: E1124 07:28:16.629237 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:28:28 crc kubenswrapper[4799]: I1124 07:28:28.628709 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:28:28 crc kubenswrapper[4799]: E1124 07:28:28.629611 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:28:39 crc kubenswrapper[4799]: I1124 07:28:39.628838 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:28:39 crc kubenswrapper[4799]: E1124 07:28:39.630127 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:28:53 crc kubenswrapper[4799]: I1124 07:28:53.628568 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:28:53 crc kubenswrapper[4799]: E1124 07:28:53.630180 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:29:05 crc kubenswrapper[4799]: I1124 07:29:05.638159 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:29:05 crc kubenswrapper[4799]: E1124 07:29:05.640369 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:29:17 crc kubenswrapper[4799]: I1124 07:29:17.629106 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:29:17 crc kubenswrapper[4799]: E1124 07:29:17.630176 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:29:29 crc kubenswrapper[4799]: I1124 07:29:29.628449 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:29:30 crc kubenswrapper[4799]: I1124 07:29:30.708471 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"a5c6678e971bdd0f47f62cf177a0fde04afd16005cad59e313430c487d2ab90d"} Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.141470 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8"] Nov 24 07:30:00 crc kubenswrapper[4799]: E1124 07:30:00.142331 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" containerName="extract-content" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.142343 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" containerName="extract-content" Nov 24 07:30:00 crc kubenswrapper[4799]: E1124 07:30:00.142453 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" containerName="registry-server" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.142463 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" containerName="registry-server" Nov 24 07:30:00 crc kubenswrapper[4799]: E1124 07:30:00.142482 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" containerName="extract-utilities" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.142496 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" containerName="extract-utilities" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.142659 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d82441-ef4d-4711-ad7c-cd7aa11d9c39" containerName="registry-server" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.143245 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.145594 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.147142 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.158424 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8"] Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.199036 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-config-volume\") pod \"collect-profiles-29399490-7wgp8\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.199431 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgkgq\" (UniqueName: \"kubernetes.io/projected/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-kube-api-access-tgkgq\") pod \"collect-profiles-29399490-7wgp8\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.199633 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-secret-volume\") pod \"collect-profiles-29399490-7wgp8\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.301369 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgkgq\" (UniqueName: \"kubernetes.io/projected/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-kube-api-access-tgkgq\") pod \"collect-profiles-29399490-7wgp8\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.301471 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-secret-volume\") pod \"collect-profiles-29399490-7wgp8\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.301559 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-config-volume\") pod \"collect-profiles-29399490-7wgp8\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.303347 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-config-volume\") pod \"collect-profiles-29399490-7wgp8\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.308638 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-secret-volume\") pod \"collect-profiles-29399490-7wgp8\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.320497 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgkgq\" (UniqueName: \"kubernetes.io/projected/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-kube-api-access-tgkgq\") pod \"collect-profiles-29399490-7wgp8\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.471908 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.950179 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8"] Nov 24 07:30:00 crc kubenswrapper[4799]: I1124 07:30:00.987376 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" event={"ID":"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b","Type":"ContainerStarted","Data":"746d50306bec1b8ec93b56ef8d426e226e68bb6e9f06c7cd2ffb0a54277a3c04"} Nov 24 07:30:01 crc kubenswrapper[4799]: I1124 07:30:01.998925 4799 generic.go:334] "Generic (PLEG): container finished" podID="6a8f2fb6-f858-45ee-b75e-9d2e5f55578b" containerID="a33694f0c2326a842ebbf4f39ce982061a35e84b590d08b0db22f212c41a4fc2" exitCode=0 Nov 24 07:30:01 crc kubenswrapper[4799]: I1124 07:30:01.999437 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" event={"ID":"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b","Type":"ContainerDied","Data":"a33694f0c2326a842ebbf4f39ce982061a35e84b590d08b0db22f212c41a4fc2"} Nov 24 07:30:03 crc kubenswrapper[4799]: I1124 07:30:03.411385 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:03 crc kubenswrapper[4799]: I1124 07:30:03.557135 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-config-volume\") pod \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " Nov 24 07:30:03 crc kubenswrapper[4799]: I1124 07:30:03.557321 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgkgq\" (UniqueName: \"kubernetes.io/projected/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-kube-api-access-tgkgq\") pod \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " Nov 24 07:30:03 crc kubenswrapper[4799]: I1124 07:30:03.557363 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-secret-volume\") pod \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\" (UID: \"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b\") " Nov 24 07:30:03 crc kubenswrapper[4799]: I1124 07:30:03.558379 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-config-volume" (OuterVolumeSpecName: "config-volume") pod "6a8f2fb6-f858-45ee-b75e-9d2e5f55578b" (UID: "6a8f2fb6-f858-45ee-b75e-9d2e5f55578b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:30:03 crc kubenswrapper[4799]: I1124 07:30:03.558524 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:03 crc kubenswrapper[4799]: I1124 07:30:03.563437 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-kube-api-access-tgkgq" (OuterVolumeSpecName: "kube-api-access-tgkgq") pod "6a8f2fb6-f858-45ee-b75e-9d2e5f55578b" (UID: "6a8f2fb6-f858-45ee-b75e-9d2e5f55578b"). InnerVolumeSpecName "kube-api-access-tgkgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:30:03 crc kubenswrapper[4799]: I1124 07:30:03.569023 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6a8f2fb6-f858-45ee-b75e-9d2e5f55578b" (UID: "6a8f2fb6-f858-45ee-b75e-9d2e5f55578b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:30:03 crc kubenswrapper[4799]: I1124 07:30:03.660507 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgkgq\" (UniqueName: \"kubernetes.io/projected/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-kube-api-access-tgkgq\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:03 crc kubenswrapper[4799]: I1124 07:30:03.660557 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:04 crc kubenswrapper[4799]: I1124 07:30:04.020272 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" event={"ID":"6a8f2fb6-f858-45ee-b75e-9d2e5f55578b","Type":"ContainerDied","Data":"746d50306bec1b8ec93b56ef8d426e226e68bb6e9f06c7cd2ffb0a54277a3c04"} Nov 24 07:30:04 crc kubenswrapper[4799]: I1124 07:30:04.020775 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="746d50306bec1b8ec93b56ef8d426e226e68bb6e9f06c7cd2ffb0a54277a3c04" Nov 24 07:30:04 crc kubenswrapper[4799]: I1124 07:30:04.020356 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8" Nov 24 07:30:04 crc kubenswrapper[4799]: I1124 07:30:04.508679 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd"] Nov 24 07:30:04 crc kubenswrapper[4799]: I1124 07:30:04.514828 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399445-2gfhd"] Nov 24 07:30:05 crc kubenswrapper[4799]: I1124 07:30:05.645013 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e116addf-dac0-42aa-9c0e-82383b214399" path="/var/lib/kubelet/pods/e116addf-dac0-42aa-9c0e-82383b214399/volumes" Nov 24 07:30:35 crc kubenswrapper[4799]: I1124 07:30:35.982975 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-grbts"] Nov 24 07:30:35 crc kubenswrapper[4799]: E1124 07:30:35.985089 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8f2fb6-f858-45ee-b75e-9d2e5f55578b" containerName="collect-profiles" Nov 24 07:30:35 crc kubenswrapper[4799]: I1124 07:30:35.985113 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8f2fb6-f858-45ee-b75e-9d2e5f55578b" containerName="collect-profiles" Nov 24 07:30:35 crc kubenswrapper[4799]: I1124 07:30:35.985378 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a8f2fb6-f858-45ee-b75e-9d2e5f55578b" containerName="collect-profiles" Nov 24 07:30:35 crc kubenswrapper[4799]: I1124 07:30:35.987395 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:35.995069 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-grbts"] Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:36.080975 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-catalog-content\") pod \"redhat-marketplace-grbts\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:36.081034 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-utilities\") pod \"redhat-marketplace-grbts\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:36.081055 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42wgx\" (UniqueName: \"kubernetes.io/projected/062ec922-c848-4214-af42-ca6853d86a79-kube-api-access-42wgx\") pod \"redhat-marketplace-grbts\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:36.182806 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-utilities\") pod \"redhat-marketplace-grbts\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:36.182912 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42wgx\" (UniqueName: \"kubernetes.io/projected/062ec922-c848-4214-af42-ca6853d86a79-kube-api-access-42wgx\") pod \"redhat-marketplace-grbts\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:36.182997 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-catalog-content\") pod \"redhat-marketplace-grbts\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:36.183313 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-utilities\") pod \"redhat-marketplace-grbts\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:36.183356 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-catalog-content\") pod \"redhat-marketplace-grbts\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:36.206282 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42wgx\" (UniqueName: \"kubernetes.io/projected/062ec922-c848-4214-af42-ca6853d86a79-kube-api-access-42wgx\") pod \"redhat-marketplace-grbts\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:36.321476 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:36 crc kubenswrapper[4799]: I1124 07:30:36.753292 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-grbts"] Nov 24 07:30:37 crc kubenswrapper[4799]: I1124 07:30:37.359542 4799 generic.go:334] "Generic (PLEG): container finished" podID="062ec922-c848-4214-af42-ca6853d86a79" containerID="18c118a3abd5657b88293085c1b3780e1d1cc11c98d89dc91209f63a9b64055d" exitCode=0 Nov 24 07:30:37 crc kubenswrapper[4799]: I1124 07:30:37.359761 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grbts" event={"ID":"062ec922-c848-4214-af42-ca6853d86a79","Type":"ContainerDied","Data":"18c118a3abd5657b88293085c1b3780e1d1cc11c98d89dc91209f63a9b64055d"} Nov 24 07:30:37 crc kubenswrapper[4799]: I1124 07:30:37.359883 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grbts" event={"ID":"062ec922-c848-4214-af42-ca6853d86a79","Type":"ContainerStarted","Data":"becbed0bd9d4e21a64398d79ac509e06cb1d5ce123b6a3474a62bd4db473a306"} Nov 24 07:30:37 crc kubenswrapper[4799]: I1124 07:30:37.365199 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:30:38 crc kubenswrapper[4799]: I1124 07:30:38.374399 4799 generic.go:334] "Generic (PLEG): container finished" podID="062ec922-c848-4214-af42-ca6853d86a79" containerID="04ca4c737579b0c308cae5084dc79063509efdc77cdccf9e97878161413538bb" exitCode=0 Nov 24 07:30:38 crc kubenswrapper[4799]: I1124 07:30:38.374493 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grbts" event={"ID":"062ec922-c848-4214-af42-ca6853d86a79","Type":"ContainerDied","Data":"04ca4c737579b0c308cae5084dc79063509efdc77cdccf9e97878161413538bb"} Nov 24 07:30:39 crc kubenswrapper[4799]: I1124 07:30:39.385317 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grbts" event={"ID":"062ec922-c848-4214-af42-ca6853d86a79","Type":"ContainerStarted","Data":"fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524"} Nov 24 07:30:39 crc kubenswrapper[4799]: I1124 07:30:39.430763 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-grbts" podStartSLOduration=2.974315724 podStartE2EDuration="4.430732097s" podCreationTimestamp="2025-11-24 07:30:35 +0000 UTC" firstStartedPulling="2025-11-24 07:30:37.364930785 +0000 UTC m=+2583.020913279" lastFinishedPulling="2025-11-24 07:30:38.821347168 +0000 UTC m=+2584.477329652" observedRunningTime="2025-11-24 07:30:39.407449876 +0000 UTC m=+2585.063432360" watchObservedRunningTime="2025-11-24 07:30:39.430732097 +0000 UTC m=+2585.086714591" Nov 24 07:30:46 crc kubenswrapper[4799]: I1124 07:30:46.322666 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:46 crc kubenswrapper[4799]: I1124 07:30:46.323364 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:46 crc kubenswrapper[4799]: I1124 07:30:46.408177 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:46 crc kubenswrapper[4799]: I1124 07:30:46.529781 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:46 crc kubenswrapper[4799]: I1124 07:30:46.657284 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-grbts"] Nov 24 07:30:48 crc kubenswrapper[4799]: I1124 07:30:48.471239 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-grbts" podUID="062ec922-c848-4214-af42-ca6853d86a79" containerName="registry-server" containerID="cri-o://fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524" gracePeriod=2 Nov 24 07:30:48 crc kubenswrapper[4799]: I1124 07:30:48.919881 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:48 crc kubenswrapper[4799]: I1124 07:30:48.976434 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-catalog-content\") pod \"062ec922-c848-4214-af42-ca6853d86a79\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " Nov 24 07:30:48 crc kubenswrapper[4799]: I1124 07:30:48.976512 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-utilities\") pod \"062ec922-c848-4214-af42-ca6853d86a79\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " Nov 24 07:30:48 crc kubenswrapper[4799]: I1124 07:30:48.976582 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42wgx\" (UniqueName: \"kubernetes.io/projected/062ec922-c848-4214-af42-ca6853d86a79-kube-api-access-42wgx\") pod \"062ec922-c848-4214-af42-ca6853d86a79\" (UID: \"062ec922-c848-4214-af42-ca6853d86a79\") " Nov 24 07:30:48 crc kubenswrapper[4799]: I1124 07:30:48.980506 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-utilities" (OuterVolumeSpecName: "utilities") pod "062ec922-c848-4214-af42-ca6853d86a79" (UID: "062ec922-c848-4214-af42-ca6853d86a79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:30:48 crc kubenswrapper[4799]: I1124 07:30:48.983113 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/062ec922-c848-4214-af42-ca6853d86a79-kube-api-access-42wgx" (OuterVolumeSpecName: "kube-api-access-42wgx") pod "062ec922-c848-4214-af42-ca6853d86a79" (UID: "062ec922-c848-4214-af42-ca6853d86a79"). InnerVolumeSpecName "kube-api-access-42wgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.005958 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "062ec922-c848-4214-af42-ca6853d86a79" (UID: "062ec922-c848-4214-af42-ca6853d86a79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.078395 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.078465 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42wgx\" (UniqueName: \"kubernetes.io/projected/062ec922-c848-4214-af42-ca6853d86a79-kube-api-access-42wgx\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.078486 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/062ec922-c848-4214-af42-ca6853d86a79-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.479255 4799 generic.go:334] "Generic (PLEG): container finished" podID="062ec922-c848-4214-af42-ca6853d86a79" containerID="fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524" exitCode=0 Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.479302 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grbts" event={"ID":"062ec922-c848-4214-af42-ca6853d86a79","Type":"ContainerDied","Data":"fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524"} Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.479333 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grbts" event={"ID":"062ec922-c848-4214-af42-ca6853d86a79","Type":"ContainerDied","Data":"becbed0bd9d4e21a64398d79ac509e06cb1d5ce123b6a3474a62bd4db473a306"} Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.479357 4799 scope.go:117] "RemoveContainer" containerID="fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.479499 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grbts" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.500513 4799 scope.go:117] "RemoveContainer" containerID="04ca4c737579b0c308cae5084dc79063509efdc77cdccf9e97878161413538bb" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.526228 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-grbts"] Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.543240 4799 scope.go:117] "RemoveContainer" containerID="18c118a3abd5657b88293085c1b3780e1d1cc11c98d89dc91209f63a9b64055d" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.554464 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-grbts"] Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.560577 4799 scope.go:117] "RemoveContainer" containerID="fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524" Nov 24 07:30:49 crc kubenswrapper[4799]: E1124 07:30:49.561027 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524\": container with ID starting with fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524 not found: ID does not exist" containerID="fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.561120 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524"} err="failed to get container status \"fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524\": rpc error: code = NotFound desc = could not find container \"fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524\": container with ID starting with fc6c6dacccdf5464a8bfd550fd00234c8a8ac0d318c0a51a724d9d3a3b434524 not found: ID does not exist" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.561200 4799 scope.go:117] "RemoveContainer" containerID="04ca4c737579b0c308cae5084dc79063509efdc77cdccf9e97878161413538bb" Nov 24 07:30:49 crc kubenswrapper[4799]: E1124 07:30:49.561486 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04ca4c737579b0c308cae5084dc79063509efdc77cdccf9e97878161413538bb\": container with ID starting with 04ca4c737579b0c308cae5084dc79063509efdc77cdccf9e97878161413538bb not found: ID does not exist" containerID="04ca4c737579b0c308cae5084dc79063509efdc77cdccf9e97878161413538bb" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.561506 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04ca4c737579b0c308cae5084dc79063509efdc77cdccf9e97878161413538bb"} err="failed to get container status \"04ca4c737579b0c308cae5084dc79063509efdc77cdccf9e97878161413538bb\": rpc error: code = NotFound desc = could not find container \"04ca4c737579b0c308cae5084dc79063509efdc77cdccf9e97878161413538bb\": container with ID starting with 04ca4c737579b0c308cae5084dc79063509efdc77cdccf9e97878161413538bb not found: ID does not exist" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.561518 4799 scope.go:117] "RemoveContainer" containerID="18c118a3abd5657b88293085c1b3780e1d1cc11c98d89dc91209f63a9b64055d" Nov 24 07:30:49 crc kubenswrapper[4799]: E1124 07:30:49.561896 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18c118a3abd5657b88293085c1b3780e1d1cc11c98d89dc91209f63a9b64055d\": container with ID starting with 18c118a3abd5657b88293085c1b3780e1d1cc11c98d89dc91209f63a9b64055d not found: ID does not exist" containerID="18c118a3abd5657b88293085c1b3780e1d1cc11c98d89dc91209f63a9b64055d" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.561978 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18c118a3abd5657b88293085c1b3780e1d1cc11c98d89dc91209f63a9b64055d"} err="failed to get container status \"18c118a3abd5657b88293085c1b3780e1d1cc11c98d89dc91209f63a9b64055d\": rpc error: code = NotFound desc = could not find container \"18c118a3abd5657b88293085c1b3780e1d1cc11c98d89dc91209f63a9b64055d\": container with ID starting with 18c118a3abd5657b88293085c1b3780e1d1cc11c98d89dc91209f63a9b64055d not found: ID does not exist" Nov 24 07:30:49 crc kubenswrapper[4799]: I1124 07:30:49.638931 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="062ec922-c848-4214-af42-ca6853d86a79" path="/var/lib/kubelet/pods/062ec922-c848-4214-af42-ca6853d86a79/volumes" Nov 24 07:30:53 crc kubenswrapper[4799]: I1124 07:30:53.078721 4799 scope.go:117] "RemoveContainer" containerID="94819cffa015d52177e06ae6f5f57848a48242776101bfdf308311b381b8e2e3" Nov 24 07:31:50 crc kubenswrapper[4799]: I1124 07:31:50.400988 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:31:50 crc kubenswrapper[4799]: I1124 07:31:50.401514 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:32:20 crc kubenswrapper[4799]: I1124 07:32:20.400235 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:32:20 crc kubenswrapper[4799]: I1124 07:32:20.400865 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:32:50 crc kubenswrapper[4799]: I1124 07:32:50.401726 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:32:50 crc kubenswrapper[4799]: I1124 07:32:50.402354 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:32:50 crc kubenswrapper[4799]: I1124 07:32:50.402412 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:32:50 crc kubenswrapper[4799]: I1124 07:32:50.403243 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a5c6678e971bdd0f47f62cf177a0fde04afd16005cad59e313430c487d2ab90d"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:32:50 crc kubenswrapper[4799]: I1124 07:32:50.403337 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://a5c6678e971bdd0f47f62cf177a0fde04afd16005cad59e313430c487d2ab90d" gracePeriod=600 Nov 24 07:32:50 crc kubenswrapper[4799]: I1124 07:32:50.670670 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="a5c6678e971bdd0f47f62cf177a0fde04afd16005cad59e313430c487d2ab90d" exitCode=0 Nov 24 07:32:50 crc kubenswrapper[4799]: I1124 07:32:50.670739 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"a5c6678e971bdd0f47f62cf177a0fde04afd16005cad59e313430c487d2ab90d"} Nov 24 07:32:50 crc kubenswrapper[4799]: I1124 07:32:50.670836 4799 scope.go:117] "RemoveContainer" containerID="a3cf5955ced558d9ab24c1bf0d2e2af4e3a7b6b7c13962c158603ec5da3df6ce" Nov 24 07:32:51 crc kubenswrapper[4799]: I1124 07:32:51.681592 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882"} Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.387486 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rcrjl"] Nov 24 07:33:30 crc kubenswrapper[4799]: E1124 07:33:30.392502 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="062ec922-c848-4214-af42-ca6853d86a79" containerName="registry-server" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.392766 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="062ec922-c848-4214-af42-ca6853d86a79" containerName="registry-server" Nov 24 07:33:30 crc kubenswrapper[4799]: E1124 07:33:30.393002 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="062ec922-c848-4214-af42-ca6853d86a79" containerName="extract-content" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.393187 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="062ec922-c848-4214-af42-ca6853d86a79" containerName="extract-content" Nov 24 07:33:30 crc kubenswrapper[4799]: E1124 07:33:30.393413 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="062ec922-c848-4214-af42-ca6853d86a79" containerName="extract-utilities" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.393609 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="062ec922-c848-4214-af42-ca6853d86a79" containerName="extract-utilities" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.394354 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="062ec922-c848-4214-af42-ca6853d86a79" containerName="registry-server" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.397129 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.418696 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rcrjl"] Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.498142 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee74762-ac3f-472e-9054-529e6b3a293f-catalog-content\") pod \"community-operators-rcrjl\" (UID: \"7ee74762-ac3f-472e-9054-529e6b3a293f\") " pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.498205 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzg8s\" (UniqueName: \"kubernetes.io/projected/7ee74762-ac3f-472e-9054-529e6b3a293f-kube-api-access-mzg8s\") pod \"community-operators-rcrjl\" (UID: \"7ee74762-ac3f-472e-9054-529e6b3a293f\") " pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.498275 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee74762-ac3f-472e-9054-529e6b3a293f-utilities\") pod \"community-operators-rcrjl\" (UID: \"7ee74762-ac3f-472e-9054-529e6b3a293f\") " pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.600164 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee74762-ac3f-472e-9054-529e6b3a293f-utilities\") pod \"community-operators-rcrjl\" (UID: \"7ee74762-ac3f-472e-9054-529e6b3a293f\") " pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.600930 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee74762-ac3f-472e-9054-529e6b3a293f-utilities\") pod \"community-operators-rcrjl\" (UID: \"7ee74762-ac3f-472e-9054-529e6b3a293f\") " pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.601246 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee74762-ac3f-472e-9054-529e6b3a293f-catalog-content\") pod \"community-operators-rcrjl\" (UID: \"7ee74762-ac3f-472e-9054-529e6b3a293f\") " pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.601653 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee74762-ac3f-472e-9054-529e6b3a293f-catalog-content\") pod \"community-operators-rcrjl\" (UID: \"7ee74762-ac3f-472e-9054-529e6b3a293f\") " pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.601749 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzg8s\" (UniqueName: \"kubernetes.io/projected/7ee74762-ac3f-472e-9054-529e6b3a293f-kube-api-access-mzg8s\") pod \"community-operators-rcrjl\" (UID: \"7ee74762-ac3f-472e-9054-529e6b3a293f\") " pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.628910 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzg8s\" (UniqueName: \"kubernetes.io/projected/7ee74762-ac3f-472e-9054-529e6b3a293f-kube-api-access-mzg8s\") pod \"community-operators-rcrjl\" (UID: \"7ee74762-ac3f-472e-9054-529e6b3a293f\") " pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:30 crc kubenswrapper[4799]: I1124 07:33:30.728319 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:31 crc kubenswrapper[4799]: I1124 07:33:31.036830 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rcrjl"] Nov 24 07:33:31 crc kubenswrapper[4799]: I1124 07:33:31.091719 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcrjl" event={"ID":"7ee74762-ac3f-472e-9054-529e6b3a293f","Type":"ContainerStarted","Data":"2b6b8b36d45177dd4ddb745ec18ecef0ac28cb31850e018732044977ca2e0cd3"} Nov 24 07:33:32 crc kubenswrapper[4799]: I1124 07:33:32.100974 4799 generic.go:334] "Generic (PLEG): container finished" podID="7ee74762-ac3f-472e-9054-529e6b3a293f" containerID="0ce724a7caa3cd262dd338cf86a08fc9a19b50fb87c0efd91464bbb9255ccb30" exitCode=0 Nov 24 07:33:32 crc kubenswrapper[4799]: I1124 07:33:32.101108 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcrjl" event={"ID":"7ee74762-ac3f-472e-9054-529e6b3a293f","Type":"ContainerDied","Data":"0ce724a7caa3cd262dd338cf86a08fc9a19b50fb87c0efd91464bbb9255ccb30"} Nov 24 07:33:36 crc kubenswrapper[4799]: I1124 07:33:36.133105 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcrjl" event={"ID":"7ee74762-ac3f-472e-9054-529e6b3a293f","Type":"ContainerStarted","Data":"4555301fccc03ec78f4034f868c4b9062015c1c11ebf89a8bb0a3b6c339bb637"} Nov 24 07:33:37 crc kubenswrapper[4799]: I1124 07:33:37.145894 4799 generic.go:334] "Generic (PLEG): container finished" podID="7ee74762-ac3f-472e-9054-529e6b3a293f" containerID="4555301fccc03ec78f4034f868c4b9062015c1c11ebf89a8bb0a3b6c339bb637" exitCode=0 Nov 24 07:33:37 crc kubenswrapper[4799]: I1124 07:33:37.146048 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcrjl" event={"ID":"7ee74762-ac3f-472e-9054-529e6b3a293f","Type":"ContainerDied","Data":"4555301fccc03ec78f4034f868c4b9062015c1c11ebf89a8bb0a3b6c339bb637"} Nov 24 07:33:38 crc kubenswrapper[4799]: I1124 07:33:38.159054 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcrjl" event={"ID":"7ee74762-ac3f-472e-9054-529e6b3a293f","Type":"ContainerStarted","Data":"49a9a1343660bdb4f2e4e64688763399389ae4144a34b984427486e9522f4728"} Nov 24 07:33:38 crc kubenswrapper[4799]: I1124 07:33:38.205978 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rcrjl" podStartSLOduration=2.7453884 podStartE2EDuration="8.205950948s" podCreationTimestamp="2025-11-24 07:33:30 +0000 UTC" firstStartedPulling="2025-11-24 07:33:32.10436969 +0000 UTC m=+2757.760352164" lastFinishedPulling="2025-11-24 07:33:37.564932198 +0000 UTC m=+2763.220914712" observedRunningTime="2025-11-24 07:33:38.192386251 +0000 UTC m=+2763.848368735" watchObservedRunningTime="2025-11-24 07:33:38.205950948 +0000 UTC m=+2763.861933442" Nov 24 07:33:40 crc kubenswrapper[4799]: I1124 07:33:40.728770 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:40 crc kubenswrapper[4799]: I1124 07:33:40.729272 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:40 crc kubenswrapper[4799]: I1124 07:33:40.810185 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:50 crc kubenswrapper[4799]: I1124 07:33:50.822544 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rcrjl" Nov 24 07:33:50 crc kubenswrapper[4799]: I1124 07:33:50.944046 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rcrjl"] Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.010094 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9spwl"] Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.010447 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9spwl" podUID="1282ec5d-e64d-463c-b5b3-15f41587f498" containerName="registry-server" containerID="cri-o://94126a15a7e912bac3e6b31ea4c2ced1f6c1dc91b7493ef32a31b262c15a8f90" gracePeriod=2 Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.290140 4799 generic.go:334] "Generic (PLEG): container finished" podID="1282ec5d-e64d-463c-b5b3-15f41587f498" containerID="94126a15a7e912bac3e6b31ea4c2ced1f6c1dc91b7493ef32a31b262c15a8f90" exitCode=0 Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.291045 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9spwl" event={"ID":"1282ec5d-e64d-463c-b5b3-15f41587f498","Type":"ContainerDied","Data":"94126a15a7e912bac3e6b31ea4c2ced1f6c1dc91b7493ef32a31b262c15a8f90"} Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.456293 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9spwl" Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.625557 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skgjt\" (UniqueName: \"kubernetes.io/projected/1282ec5d-e64d-463c-b5b3-15f41587f498-kube-api-access-skgjt\") pod \"1282ec5d-e64d-463c-b5b3-15f41587f498\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.625668 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-utilities\") pod \"1282ec5d-e64d-463c-b5b3-15f41587f498\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.625711 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-catalog-content\") pod \"1282ec5d-e64d-463c-b5b3-15f41587f498\" (UID: \"1282ec5d-e64d-463c-b5b3-15f41587f498\") " Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.626751 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-utilities" (OuterVolumeSpecName: "utilities") pod "1282ec5d-e64d-463c-b5b3-15f41587f498" (UID: "1282ec5d-e64d-463c-b5b3-15f41587f498"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.631958 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1282ec5d-e64d-463c-b5b3-15f41587f498-kube-api-access-skgjt" (OuterVolumeSpecName: "kube-api-access-skgjt") pod "1282ec5d-e64d-463c-b5b3-15f41587f498" (UID: "1282ec5d-e64d-463c-b5b3-15f41587f498"). InnerVolumeSpecName "kube-api-access-skgjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.695803 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1282ec5d-e64d-463c-b5b3-15f41587f498" (UID: "1282ec5d-e64d-463c-b5b3-15f41587f498"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.727721 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skgjt\" (UniqueName: \"kubernetes.io/projected/1282ec5d-e64d-463c-b5b3-15f41587f498-kube-api-access-skgjt\") on node \"crc\" DevicePath \"\"" Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.727756 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:33:51 crc kubenswrapper[4799]: I1124 07:33:51.727768 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1282ec5d-e64d-463c-b5b3-15f41587f498-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:33:52 crc kubenswrapper[4799]: I1124 07:33:52.304004 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9spwl" event={"ID":"1282ec5d-e64d-463c-b5b3-15f41587f498","Type":"ContainerDied","Data":"5cebebcef9944ef817943117d07c981215a71c1a8c4bf6f78ca08d4dcd809a5e"} Nov 24 07:33:52 crc kubenswrapper[4799]: I1124 07:33:52.304295 4799 scope.go:117] "RemoveContainer" containerID="94126a15a7e912bac3e6b31ea4c2ced1f6c1dc91b7493ef32a31b262c15a8f90" Nov 24 07:33:52 crc kubenswrapper[4799]: I1124 07:33:52.304116 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9spwl" Nov 24 07:33:52 crc kubenswrapper[4799]: I1124 07:33:52.346793 4799 scope.go:117] "RemoveContainer" containerID="a66a7fda1ec8146d70589bd8eb6e793b7edeed60e7b21ba935ad2d36a29e4abb" Nov 24 07:33:52 crc kubenswrapper[4799]: I1124 07:33:52.354714 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9spwl"] Nov 24 07:33:52 crc kubenswrapper[4799]: I1124 07:33:52.361036 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9spwl"] Nov 24 07:33:52 crc kubenswrapper[4799]: I1124 07:33:52.369332 4799 scope.go:117] "RemoveContainer" containerID="d142866ab8960171954d38ac87bbcf6697a0fe5151d50fe1a0460987cf823424" Nov 24 07:33:52 crc kubenswrapper[4799]: E1124 07:33:52.395558 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1282ec5d_e64d_463c_b5b3_15f41587f498.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1282ec5d_e64d_463c_b5b3_15f41587f498.slice/crio-5cebebcef9944ef817943117d07c981215a71c1a8c4bf6f78ca08d4dcd809a5e\": RecentStats: unable to find data in memory cache]" Nov 24 07:33:53 crc kubenswrapper[4799]: I1124 07:33:53.636794 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1282ec5d-e64d-463c-b5b3-15f41587f498" path="/var/lib/kubelet/pods/1282ec5d-e64d-463c-b5b3-15f41587f498/volumes" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.309133 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kssmh"] Nov 24 07:34:31 crc kubenswrapper[4799]: E1124 07:34:31.309914 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1282ec5d-e64d-463c-b5b3-15f41587f498" containerName="registry-server" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.309926 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1282ec5d-e64d-463c-b5b3-15f41587f498" containerName="registry-server" Nov 24 07:34:31 crc kubenswrapper[4799]: E1124 07:34:31.309940 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1282ec5d-e64d-463c-b5b3-15f41587f498" containerName="extract-utilities" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.309947 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1282ec5d-e64d-463c-b5b3-15f41587f498" containerName="extract-utilities" Nov 24 07:34:31 crc kubenswrapper[4799]: E1124 07:34:31.309956 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1282ec5d-e64d-463c-b5b3-15f41587f498" containerName="extract-content" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.309962 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1282ec5d-e64d-463c-b5b3-15f41587f498" containerName="extract-content" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.310103 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1282ec5d-e64d-463c-b5b3-15f41587f498" containerName="registry-server" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.311076 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.315114 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kssmh"] Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.366053 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-catalog-content\") pod \"redhat-operators-kssmh\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.366192 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24jc9\" (UniqueName: \"kubernetes.io/projected/e3c00960-59e9-4288-a4cd-38fd635552e7-kube-api-access-24jc9\") pod \"redhat-operators-kssmh\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.366319 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-utilities\") pod \"redhat-operators-kssmh\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.467298 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-utilities\") pod \"redhat-operators-kssmh\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.467349 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-catalog-content\") pod \"redhat-operators-kssmh\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.467408 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24jc9\" (UniqueName: \"kubernetes.io/projected/e3c00960-59e9-4288-a4cd-38fd635552e7-kube-api-access-24jc9\") pod \"redhat-operators-kssmh\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.468474 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-utilities\") pod \"redhat-operators-kssmh\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.468541 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-catalog-content\") pod \"redhat-operators-kssmh\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.490364 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24jc9\" (UniqueName: \"kubernetes.io/projected/e3c00960-59e9-4288-a4cd-38fd635552e7-kube-api-access-24jc9\") pod \"redhat-operators-kssmh\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:31 crc kubenswrapper[4799]: I1124 07:34:31.647149 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:32 crc kubenswrapper[4799]: I1124 07:34:32.107694 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kssmh"] Nov 24 07:34:32 crc kubenswrapper[4799]: I1124 07:34:32.697002 4799 generic.go:334] "Generic (PLEG): container finished" podID="e3c00960-59e9-4288-a4cd-38fd635552e7" containerID="b43d5c9d0d7334aede5264fba202487cb36a0e367ee6ab91ff5b0df6f2d04458" exitCode=0 Nov 24 07:34:32 crc kubenswrapper[4799]: I1124 07:34:32.697209 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kssmh" event={"ID":"e3c00960-59e9-4288-a4cd-38fd635552e7","Type":"ContainerDied","Data":"b43d5c9d0d7334aede5264fba202487cb36a0e367ee6ab91ff5b0df6f2d04458"} Nov 24 07:34:32 crc kubenswrapper[4799]: I1124 07:34:32.697262 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kssmh" event={"ID":"e3c00960-59e9-4288-a4cd-38fd635552e7","Type":"ContainerStarted","Data":"4fb7cc3bc90a7070680dff8ede81ffc9e14403892606fe751dd5165ec2a84e0e"} Nov 24 07:34:40 crc kubenswrapper[4799]: I1124 07:34:40.082910 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kssmh" event={"ID":"e3c00960-59e9-4288-a4cd-38fd635552e7","Type":"ContainerStarted","Data":"163d1cd347202e161f77ab1534d2055440d37b6e2a64ab8cd534c1c6c2b2e552"} Nov 24 07:34:41 crc kubenswrapper[4799]: I1124 07:34:41.096611 4799 generic.go:334] "Generic (PLEG): container finished" podID="e3c00960-59e9-4288-a4cd-38fd635552e7" containerID="163d1cd347202e161f77ab1534d2055440d37b6e2a64ab8cd534c1c6c2b2e552" exitCode=0 Nov 24 07:34:41 crc kubenswrapper[4799]: I1124 07:34:41.096693 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kssmh" event={"ID":"e3c00960-59e9-4288-a4cd-38fd635552e7","Type":"ContainerDied","Data":"163d1cd347202e161f77ab1534d2055440d37b6e2a64ab8cd534c1c6c2b2e552"} Nov 24 07:34:42 crc kubenswrapper[4799]: I1124 07:34:42.110616 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kssmh" event={"ID":"e3c00960-59e9-4288-a4cd-38fd635552e7","Type":"ContainerStarted","Data":"ac4366c346d6c0ba28251c98afde3dfa36b15fba9d2ad3ebfa646406c85f15be"} Nov 24 07:34:42 crc kubenswrapper[4799]: I1124 07:34:42.137279 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kssmh" podStartSLOduration=2.223497622 podStartE2EDuration="11.137254716s" podCreationTimestamp="2025-11-24 07:34:31 +0000 UTC" firstStartedPulling="2025-11-24 07:34:32.698492333 +0000 UTC m=+2818.354474807" lastFinishedPulling="2025-11-24 07:34:41.612249387 +0000 UTC m=+2827.268231901" observedRunningTime="2025-11-24 07:34:42.132342056 +0000 UTC m=+2827.788324550" watchObservedRunningTime="2025-11-24 07:34:42.137254716 +0000 UTC m=+2827.793237200" Nov 24 07:34:50 crc kubenswrapper[4799]: I1124 07:34:50.400910 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:34:50 crc kubenswrapper[4799]: I1124 07:34:50.401433 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:34:51 crc kubenswrapper[4799]: I1124 07:34:51.647839 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:51 crc kubenswrapper[4799]: I1124 07:34:51.647957 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:51 crc kubenswrapper[4799]: I1124 07:34:51.722414 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:52 crc kubenswrapper[4799]: I1124 07:34:52.240598 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 07:34:52 crc kubenswrapper[4799]: I1124 07:34:52.330593 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kssmh"] Nov 24 07:34:52 crc kubenswrapper[4799]: I1124 07:34:52.352739 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4wd9p"] Nov 24 07:34:52 crc kubenswrapper[4799]: I1124 07:34:52.353109 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4wd9p" podUID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" containerName="registry-server" containerID="cri-o://ec4e4bc1527a4dbc22f991c1ed8d40c3f2b359f4bfbb1fe649149bf37ab7f83b" gracePeriod=2 Nov 24 07:34:53 crc kubenswrapper[4799]: I1124 07:34:53.212393 4799 generic.go:334] "Generic (PLEG): container finished" podID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" containerID="ec4e4bc1527a4dbc22f991c1ed8d40c3f2b359f4bfbb1fe649149bf37ab7f83b" exitCode=0 Nov 24 07:34:53 crc kubenswrapper[4799]: I1124 07:34:53.212505 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wd9p" event={"ID":"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624","Type":"ContainerDied","Data":"ec4e4bc1527a4dbc22f991c1ed8d40c3f2b359f4bfbb1fe649149bf37ab7f83b"} Nov 24 07:34:54 crc kubenswrapper[4799]: I1124 07:34:54.629554 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 07:34:54 crc kubenswrapper[4799]: I1124 07:34:54.818764 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-catalog-content\") pod \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " Nov 24 07:34:54 crc kubenswrapper[4799]: I1124 07:34:54.818898 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfjm4\" (UniqueName: \"kubernetes.io/projected/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-kube-api-access-lfjm4\") pod \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " Nov 24 07:34:54 crc kubenswrapper[4799]: I1124 07:34:54.818980 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-utilities\") pod \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\" (UID: \"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624\") " Nov 24 07:34:54 crc kubenswrapper[4799]: I1124 07:34:54.819468 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-utilities" (OuterVolumeSpecName: "utilities") pod "e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" (UID: "e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:34:54 crc kubenswrapper[4799]: I1124 07:34:54.828166 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-kube-api-access-lfjm4" (OuterVolumeSpecName: "kube-api-access-lfjm4") pod "e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" (UID: "e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624"). InnerVolumeSpecName "kube-api-access-lfjm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:34:54 crc kubenswrapper[4799]: I1124 07:34:54.905831 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" (UID: "e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:34:54 crc kubenswrapper[4799]: I1124 07:34:54.920354 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:34:54 crc kubenswrapper[4799]: I1124 07:34:54.920404 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfjm4\" (UniqueName: \"kubernetes.io/projected/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-kube-api-access-lfjm4\") on node \"crc\" DevicePath \"\"" Nov 24 07:34:54 crc kubenswrapper[4799]: I1124 07:34:54.920420 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:34:55 crc kubenswrapper[4799]: I1124 07:34:55.227005 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wd9p" event={"ID":"e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624","Type":"ContainerDied","Data":"d9c21babf0e7af128a8de425d8d68dd526f65d3c5b32e60bb184e4186ddf6b31"} Nov 24 07:34:55 crc kubenswrapper[4799]: I1124 07:34:55.227067 4799 scope.go:117] "RemoveContainer" containerID="ec4e4bc1527a4dbc22f991c1ed8d40c3f2b359f4bfbb1fe649149bf37ab7f83b" Nov 24 07:34:55 crc kubenswrapper[4799]: I1124 07:34:55.227111 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wd9p" Nov 24 07:34:55 crc kubenswrapper[4799]: I1124 07:34:55.248673 4799 scope.go:117] "RemoveContainer" containerID="900bb6ba8a0319762fe7b1b5c9b3f72e1bc6a74964d4915427b7ec55ba29ed95" Nov 24 07:34:55 crc kubenswrapper[4799]: I1124 07:34:55.269258 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4wd9p"] Nov 24 07:34:55 crc kubenswrapper[4799]: I1124 07:34:55.275524 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4wd9p"] Nov 24 07:34:55 crc kubenswrapper[4799]: I1124 07:34:55.280792 4799 scope.go:117] "RemoveContainer" containerID="3d4549c36635a0a490a1efb8b0bea23c5a2b35e2f2b47bb44afc033260334f85" Nov 24 07:34:55 crc kubenswrapper[4799]: I1124 07:34:55.647517 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" path="/var/lib/kubelet/pods/e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624/volumes" Nov 24 07:35:20 crc kubenswrapper[4799]: I1124 07:35:20.401440 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:35:20 crc kubenswrapper[4799]: I1124 07:35:20.402379 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:35:50 crc kubenswrapper[4799]: I1124 07:35:50.400780 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:35:50 crc kubenswrapper[4799]: I1124 07:35:50.401623 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:35:50 crc kubenswrapper[4799]: I1124 07:35:50.401691 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:35:50 crc kubenswrapper[4799]: I1124 07:35:50.402579 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:35:50 crc kubenswrapper[4799]: I1124 07:35:50.402680 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" gracePeriod=600 Nov 24 07:35:50 crc kubenswrapper[4799]: E1124 07:35:50.533386 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:35:50 crc kubenswrapper[4799]: I1124 07:35:50.739531 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" exitCode=0 Nov 24 07:35:50 crc kubenswrapper[4799]: I1124 07:35:50.739634 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882"} Nov 24 07:35:50 crc kubenswrapper[4799]: I1124 07:35:50.739677 4799 scope.go:117] "RemoveContainer" containerID="a5c6678e971bdd0f47f62cf177a0fde04afd16005cad59e313430c487d2ab90d" Nov 24 07:35:50 crc kubenswrapper[4799]: I1124 07:35:50.740434 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:35:50 crc kubenswrapper[4799]: E1124 07:35:50.740906 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:36:03 crc kubenswrapper[4799]: I1124 07:36:03.629091 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:36:03 crc kubenswrapper[4799]: E1124 07:36:03.630154 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:36:17 crc kubenswrapper[4799]: I1124 07:36:17.631547 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:36:17 crc kubenswrapper[4799]: E1124 07:36:17.632664 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:36:30 crc kubenswrapper[4799]: I1124 07:36:30.628561 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:36:30 crc kubenswrapper[4799]: E1124 07:36:30.630439 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.100346 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p2svm"] Nov 24 07:36:36 crc kubenswrapper[4799]: E1124 07:36:36.101296 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" containerName="extract-content" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.101311 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" containerName="extract-content" Nov 24 07:36:36 crc kubenswrapper[4799]: E1124 07:36:36.101328 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" containerName="extract-utilities" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.101336 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" containerName="extract-utilities" Nov 24 07:36:36 crc kubenswrapper[4799]: E1124 07:36:36.101360 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" containerName="registry-server" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.101368 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" containerName="registry-server" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.101529 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e48bbcd3-6bb0-4f7d-9bac-46dd1a70e624" containerName="registry-server" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.102632 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.120697 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p2svm"] Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.149159 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-utilities\") pod \"certified-operators-p2svm\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.149225 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-catalog-content\") pod \"certified-operators-p2svm\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.149262 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w68pg\" (UniqueName: \"kubernetes.io/projected/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-kube-api-access-w68pg\") pod \"certified-operators-p2svm\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.250244 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-utilities\") pod \"certified-operators-p2svm\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.250307 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-catalog-content\") pod \"certified-operators-p2svm\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.250340 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w68pg\" (UniqueName: \"kubernetes.io/projected/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-kube-api-access-w68pg\") pod \"certified-operators-p2svm\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.250971 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-utilities\") pod \"certified-operators-p2svm\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.251105 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-catalog-content\") pod \"certified-operators-p2svm\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.273875 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w68pg\" (UniqueName: \"kubernetes.io/projected/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-kube-api-access-w68pg\") pod \"certified-operators-p2svm\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.426655 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:36 crc kubenswrapper[4799]: I1124 07:36:36.681766 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p2svm"] Nov 24 07:36:36 crc kubenswrapper[4799]: W1124 07:36:36.690837 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e20b0f5_f6f2_46a6_9485_62c74553d9e8.slice/crio-f89247d4825a38f202fb9604e1d1fef721fa237f6a75032a6845e04d426595fd WatchSource:0}: Error finding container f89247d4825a38f202fb9604e1d1fef721fa237f6a75032a6845e04d426595fd: Status 404 returned error can't find the container with id f89247d4825a38f202fb9604e1d1fef721fa237f6a75032a6845e04d426595fd Nov 24 07:36:37 crc kubenswrapper[4799]: I1124 07:36:37.177758 4799 generic.go:334] "Generic (PLEG): container finished" podID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" containerID="88867127258d9ca83f2c528b752451f6b421ceb1a0da378dba0d38c75f73bcb9" exitCode=0 Nov 24 07:36:37 crc kubenswrapper[4799]: I1124 07:36:37.177825 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2svm" event={"ID":"2e20b0f5-f6f2-46a6-9485-62c74553d9e8","Type":"ContainerDied","Data":"88867127258d9ca83f2c528b752451f6b421ceb1a0da378dba0d38c75f73bcb9"} Nov 24 07:36:37 crc kubenswrapper[4799]: I1124 07:36:37.177893 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2svm" event={"ID":"2e20b0f5-f6f2-46a6-9485-62c74553d9e8","Type":"ContainerStarted","Data":"f89247d4825a38f202fb9604e1d1fef721fa237f6a75032a6845e04d426595fd"} Nov 24 07:36:37 crc kubenswrapper[4799]: I1124 07:36:37.180933 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:36:39 crc kubenswrapper[4799]: I1124 07:36:39.198043 4799 generic.go:334] "Generic (PLEG): container finished" podID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" containerID="8a34afc4b3db43bbddbc6d6e52974f0f82c096f19b0a2722e4c4ebb92e4a04bb" exitCode=0 Nov 24 07:36:39 crc kubenswrapper[4799]: I1124 07:36:39.198100 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2svm" event={"ID":"2e20b0f5-f6f2-46a6-9485-62c74553d9e8","Type":"ContainerDied","Data":"8a34afc4b3db43bbddbc6d6e52974f0f82c096f19b0a2722e4c4ebb92e4a04bb"} Nov 24 07:36:40 crc kubenswrapper[4799]: I1124 07:36:40.207822 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2svm" event={"ID":"2e20b0f5-f6f2-46a6-9485-62c74553d9e8","Type":"ContainerStarted","Data":"eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659"} Nov 24 07:36:40 crc kubenswrapper[4799]: I1124 07:36:40.229210 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p2svm" podStartSLOduration=1.737379458 podStartE2EDuration="4.229190394s" podCreationTimestamp="2025-11-24 07:36:36 +0000 UTC" firstStartedPulling="2025-11-24 07:36:37.180135699 +0000 UTC m=+2942.836118213" lastFinishedPulling="2025-11-24 07:36:39.671946635 +0000 UTC m=+2945.327929149" observedRunningTime="2025-11-24 07:36:40.228862835 +0000 UTC m=+2945.884845309" watchObservedRunningTime="2025-11-24 07:36:40.229190394 +0000 UTC m=+2945.885172878" Nov 24 07:36:43 crc kubenswrapper[4799]: I1124 07:36:43.628075 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:36:43 crc kubenswrapper[4799]: E1124 07:36:43.628675 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:36:46 crc kubenswrapper[4799]: I1124 07:36:46.427950 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:46 crc kubenswrapper[4799]: I1124 07:36:46.428358 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:46 crc kubenswrapper[4799]: I1124 07:36:46.505670 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:47 crc kubenswrapper[4799]: I1124 07:36:47.350709 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:47 crc kubenswrapper[4799]: I1124 07:36:47.433399 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p2svm"] Nov 24 07:36:49 crc kubenswrapper[4799]: I1124 07:36:49.292700 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p2svm" podUID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" containerName="registry-server" containerID="cri-o://eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659" gracePeriod=2 Nov 24 07:36:49 crc kubenswrapper[4799]: I1124 07:36:49.752233 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:49 crc kubenswrapper[4799]: I1124 07:36:49.902127 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-catalog-content\") pod \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " Nov 24 07:36:49 crc kubenswrapper[4799]: I1124 07:36:49.902270 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-utilities\") pod \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " Nov 24 07:36:49 crc kubenswrapper[4799]: I1124 07:36:49.902337 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w68pg\" (UniqueName: \"kubernetes.io/projected/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-kube-api-access-w68pg\") pod \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\" (UID: \"2e20b0f5-f6f2-46a6-9485-62c74553d9e8\") " Nov 24 07:36:49 crc kubenswrapper[4799]: I1124 07:36:49.903728 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-utilities" (OuterVolumeSpecName: "utilities") pod "2e20b0f5-f6f2-46a6-9485-62c74553d9e8" (UID: "2e20b0f5-f6f2-46a6-9485-62c74553d9e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:36:49 crc kubenswrapper[4799]: I1124 07:36:49.908510 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-kube-api-access-w68pg" (OuterVolumeSpecName: "kube-api-access-w68pg") pod "2e20b0f5-f6f2-46a6-9485-62c74553d9e8" (UID: "2e20b0f5-f6f2-46a6-9485-62c74553d9e8"). InnerVolumeSpecName "kube-api-access-w68pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:36:49 crc kubenswrapper[4799]: I1124 07:36:49.950501 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e20b0f5-f6f2-46a6-9485-62c74553d9e8" (UID: "2e20b0f5-f6f2-46a6-9485-62c74553d9e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.004595 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.004649 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w68pg\" (UniqueName: \"kubernetes.io/projected/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-kube-api-access-w68pg\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.004675 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e20b0f5-f6f2-46a6-9485-62c74553d9e8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.306074 4799 generic.go:334] "Generic (PLEG): container finished" podID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" containerID="eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659" exitCode=0 Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.306148 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2svm" event={"ID":"2e20b0f5-f6f2-46a6-9485-62c74553d9e8","Type":"ContainerDied","Data":"eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659"} Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.306229 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2svm" event={"ID":"2e20b0f5-f6f2-46a6-9485-62c74553d9e8","Type":"ContainerDied","Data":"f89247d4825a38f202fb9604e1d1fef721fa237f6a75032a6845e04d426595fd"} Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.306271 4799 scope.go:117] "RemoveContainer" containerID="eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.306175 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2svm" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.340822 4799 scope.go:117] "RemoveContainer" containerID="8a34afc4b3db43bbddbc6d6e52974f0f82c096f19b0a2722e4c4ebb92e4a04bb" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.358279 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p2svm"] Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.368869 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p2svm"] Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.375248 4799 scope.go:117] "RemoveContainer" containerID="88867127258d9ca83f2c528b752451f6b421ceb1a0da378dba0d38c75f73bcb9" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.403183 4799 scope.go:117] "RemoveContainer" containerID="eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659" Nov 24 07:36:50 crc kubenswrapper[4799]: E1124 07:36:50.403695 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659\": container with ID starting with eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659 not found: ID does not exist" containerID="eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.403731 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659"} err="failed to get container status \"eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659\": rpc error: code = NotFound desc = could not find container \"eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659\": container with ID starting with eca7f14eb0899391d2d5432217d311c4a25a0123f952ea6f7060f92b189bf659 not found: ID does not exist" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.403759 4799 scope.go:117] "RemoveContainer" containerID="8a34afc4b3db43bbddbc6d6e52974f0f82c096f19b0a2722e4c4ebb92e4a04bb" Nov 24 07:36:50 crc kubenswrapper[4799]: E1124 07:36:50.404166 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a34afc4b3db43bbddbc6d6e52974f0f82c096f19b0a2722e4c4ebb92e4a04bb\": container with ID starting with 8a34afc4b3db43bbddbc6d6e52974f0f82c096f19b0a2722e4c4ebb92e4a04bb not found: ID does not exist" containerID="8a34afc4b3db43bbddbc6d6e52974f0f82c096f19b0a2722e4c4ebb92e4a04bb" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.404344 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a34afc4b3db43bbddbc6d6e52974f0f82c096f19b0a2722e4c4ebb92e4a04bb"} err="failed to get container status \"8a34afc4b3db43bbddbc6d6e52974f0f82c096f19b0a2722e4c4ebb92e4a04bb\": rpc error: code = NotFound desc = could not find container \"8a34afc4b3db43bbddbc6d6e52974f0f82c096f19b0a2722e4c4ebb92e4a04bb\": container with ID starting with 8a34afc4b3db43bbddbc6d6e52974f0f82c096f19b0a2722e4c4ebb92e4a04bb not found: ID does not exist" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.404491 4799 scope.go:117] "RemoveContainer" containerID="88867127258d9ca83f2c528b752451f6b421ceb1a0da378dba0d38c75f73bcb9" Nov 24 07:36:50 crc kubenswrapper[4799]: E1124 07:36:50.404986 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88867127258d9ca83f2c528b752451f6b421ceb1a0da378dba0d38c75f73bcb9\": container with ID starting with 88867127258d9ca83f2c528b752451f6b421ceb1a0da378dba0d38c75f73bcb9 not found: ID does not exist" containerID="88867127258d9ca83f2c528b752451f6b421ceb1a0da378dba0d38c75f73bcb9" Nov 24 07:36:50 crc kubenswrapper[4799]: I1124 07:36:50.405038 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88867127258d9ca83f2c528b752451f6b421ceb1a0da378dba0d38c75f73bcb9"} err="failed to get container status \"88867127258d9ca83f2c528b752451f6b421ceb1a0da378dba0d38c75f73bcb9\": rpc error: code = NotFound desc = could not find container \"88867127258d9ca83f2c528b752451f6b421ceb1a0da378dba0d38c75f73bcb9\": container with ID starting with 88867127258d9ca83f2c528b752451f6b421ceb1a0da378dba0d38c75f73bcb9 not found: ID does not exist" Nov 24 07:36:51 crc kubenswrapper[4799]: I1124 07:36:51.642773 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" path="/var/lib/kubelet/pods/2e20b0f5-f6f2-46a6-9485-62c74553d9e8/volumes" Nov 24 07:36:55 crc kubenswrapper[4799]: I1124 07:36:55.636772 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:36:55 crc kubenswrapper[4799]: E1124 07:36:55.637809 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:37:10 crc kubenswrapper[4799]: I1124 07:37:10.628679 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:37:10 crc kubenswrapper[4799]: E1124 07:37:10.629813 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:37:25 crc kubenswrapper[4799]: I1124 07:37:25.639670 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:37:25 crc kubenswrapper[4799]: E1124 07:37:25.641393 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:37:37 crc kubenswrapper[4799]: I1124 07:37:37.627827 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:37:37 crc kubenswrapper[4799]: E1124 07:37:37.629777 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:37:52 crc kubenswrapper[4799]: I1124 07:37:52.628451 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:37:52 crc kubenswrapper[4799]: E1124 07:37:52.629669 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:38:04 crc kubenswrapper[4799]: I1124 07:38:04.628422 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:38:04 crc kubenswrapper[4799]: E1124 07:38:04.629685 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:38:16 crc kubenswrapper[4799]: I1124 07:38:16.628523 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:38:16 crc kubenswrapper[4799]: E1124 07:38:16.629692 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:38:31 crc kubenswrapper[4799]: I1124 07:38:31.628930 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:38:31 crc kubenswrapper[4799]: E1124 07:38:31.632755 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:38:42 crc kubenswrapper[4799]: I1124 07:38:42.628172 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:38:42 crc kubenswrapper[4799]: E1124 07:38:42.628807 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:38:57 crc kubenswrapper[4799]: I1124 07:38:57.628907 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:38:57 crc kubenswrapper[4799]: E1124 07:38:57.629990 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:39:11 crc kubenswrapper[4799]: I1124 07:39:11.629102 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:39:11 crc kubenswrapper[4799]: E1124 07:39:11.630131 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:39:24 crc kubenswrapper[4799]: I1124 07:39:24.627941 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:39:24 crc kubenswrapper[4799]: E1124 07:39:24.628598 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:39:37 crc kubenswrapper[4799]: I1124 07:39:37.628359 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:39:37 crc kubenswrapper[4799]: E1124 07:39:37.629619 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:39:51 crc kubenswrapper[4799]: I1124 07:39:51.629083 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:39:51 crc kubenswrapper[4799]: E1124 07:39:51.630319 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:40:06 crc kubenswrapper[4799]: I1124 07:40:06.628724 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:40:06 crc kubenswrapper[4799]: E1124 07:40:06.629458 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:40:21 crc kubenswrapper[4799]: I1124 07:40:21.628634 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:40:21 crc kubenswrapper[4799]: E1124 07:40:21.629642 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:40:35 crc kubenswrapper[4799]: I1124 07:40:35.636378 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:40:35 crc kubenswrapper[4799]: E1124 07:40:35.637393 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:40:48 crc kubenswrapper[4799]: I1124 07:40:48.629419 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:40:48 crc kubenswrapper[4799]: E1124 07:40:48.630609 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:41:00 crc kubenswrapper[4799]: I1124 07:41:00.628919 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:41:00 crc kubenswrapper[4799]: I1124 07:41:00.877547 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"80a5fa14424f6f54364c175b542535d6d16c363682f8b22f6fe58a3c63c8ac33"} Nov 24 07:41:28 crc kubenswrapper[4799]: I1124 07:41:28.966690 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ml8"] Nov 24 07:41:28 crc kubenswrapper[4799]: E1124 07:41:28.967969 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" containerName="extract-utilities" Nov 24 07:41:28 crc kubenswrapper[4799]: I1124 07:41:28.967998 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" containerName="extract-utilities" Nov 24 07:41:28 crc kubenswrapper[4799]: E1124 07:41:28.968049 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" containerName="extract-content" Nov 24 07:41:28 crc kubenswrapper[4799]: I1124 07:41:28.968067 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" containerName="extract-content" Nov 24 07:41:28 crc kubenswrapper[4799]: E1124 07:41:28.968088 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" containerName="registry-server" Nov 24 07:41:28 crc kubenswrapper[4799]: I1124 07:41:28.968104 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" containerName="registry-server" Nov 24 07:41:28 crc kubenswrapper[4799]: I1124 07:41:28.968446 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e20b0f5-f6f2-46a6-9485-62c74553d9e8" containerName="registry-server" Nov 24 07:41:28 crc kubenswrapper[4799]: I1124 07:41:28.970502 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:28 crc kubenswrapper[4799]: I1124 07:41:28.986450 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4m4w\" (UniqueName: \"kubernetes.io/projected/dcdca964-4f09-4026-926f-0abf69251368-kube-api-access-f4m4w\") pod \"redhat-marketplace-h4ml8\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:28 crc kubenswrapper[4799]: I1124 07:41:28.986546 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-utilities\") pod \"redhat-marketplace-h4ml8\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:28 crc kubenswrapper[4799]: I1124 07:41:28.986665 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-catalog-content\") pod \"redhat-marketplace-h4ml8\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:28 crc kubenswrapper[4799]: I1124 07:41:28.995282 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ml8"] Nov 24 07:41:29 crc kubenswrapper[4799]: I1124 07:41:29.087861 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4m4w\" (UniqueName: \"kubernetes.io/projected/dcdca964-4f09-4026-926f-0abf69251368-kube-api-access-f4m4w\") pod \"redhat-marketplace-h4ml8\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:29 crc kubenswrapper[4799]: I1124 07:41:29.087914 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-utilities\") pod \"redhat-marketplace-h4ml8\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:29 crc kubenswrapper[4799]: I1124 07:41:29.087957 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-catalog-content\") pod \"redhat-marketplace-h4ml8\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:29 crc kubenswrapper[4799]: I1124 07:41:29.088398 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-catalog-content\") pod \"redhat-marketplace-h4ml8\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:29 crc kubenswrapper[4799]: I1124 07:41:29.088544 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-utilities\") pod \"redhat-marketplace-h4ml8\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:29 crc kubenswrapper[4799]: I1124 07:41:29.113035 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4m4w\" (UniqueName: \"kubernetes.io/projected/dcdca964-4f09-4026-926f-0abf69251368-kube-api-access-f4m4w\") pod \"redhat-marketplace-h4ml8\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:29 crc kubenswrapper[4799]: I1124 07:41:29.350867 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:29 crc kubenswrapper[4799]: I1124 07:41:29.596232 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ml8"] Nov 24 07:41:30 crc kubenswrapper[4799]: I1124 07:41:30.181425 4799 generic.go:334] "Generic (PLEG): container finished" podID="dcdca964-4f09-4026-926f-0abf69251368" containerID="8556a5ae88ceab5092ed484e0b2ac079b80b92ad817c9d304102039eecc19b4d" exitCode=0 Nov 24 07:41:30 crc kubenswrapper[4799]: I1124 07:41:30.181609 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ml8" event={"ID":"dcdca964-4f09-4026-926f-0abf69251368","Type":"ContainerDied","Data":"8556a5ae88ceab5092ed484e0b2ac079b80b92ad817c9d304102039eecc19b4d"} Nov 24 07:41:30 crc kubenswrapper[4799]: I1124 07:41:30.181942 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ml8" event={"ID":"dcdca964-4f09-4026-926f-0abf69251368","Type":"ContainerStarted","Data":"c4b836ef487e70dd404f9a38bcb11fda7425ab92374ea7c9ee711bad4f7d43e6"} Nov 24 07:41:31 crc kubenswrapper[4799]: I1124 07:41:31.192787 4799 generic.go:334] "Generic (PLEG): container finished" podID="dcdca964-4f09-4026-926f-0abf69251368" containerID="87098c8ccfc2041a8fb43759cde2c2d463d48e0509784ddba5a2b4b774385621" exitCode=0 Nov 24 07:41:31 crc kubenswrapper[4799]: I1124 07:41:31.192920 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ml8" event={"ID":"dcdca964-4f09-4026-926f-0abf69251368","Type":"ContainerDied","Data":"87098c8ccfc2041a8fb43759cde2c2d463d48e0509784ddba5a2b4b774385621"} Nov 24 07:41:32 crc kubenswrapper[4799]: I1124 07:41:32.206281 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ml8" event={"ID":"dcdca964-4f09-4026-926f-0abf69251368","Type":"ContainerStarted","Data":"e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712"} Nov 24 07:41:32 crc kubenswrapper[4799]: I1124 07:41:32.238746 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h4ml8" podStartSLOduration=2.839309183 podStartE2EDuration="4.23872438s" podCreationTimestamp="2025-11-24 07:41:28 +0000 UTC" firstStartedPulling="2025-11-24 07:41:30.185200372 +0000 UTC m=+3235.841182886" lastFinishedPulling="2025-11-24 07:41:31.584615599 +0000 UTC m=+3237.240598083" observedRunningTime="2025-11-24 07:41:32.237942608 +0000 UTC m=+3237.893925122" watchObservedRunningTime="2025-11-24 07:41:32.23872438 +0000 UTC m=+3237.894706874" Nov 24 07:41:39 crc kubenswrapper[4799]: I1124 07:41:39.351390 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:39 crc kubenswrapper[4799]: I1124 07:41:39.352314 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:39 crc kubenswrapper[4799]: I1124 07:41:39.429265 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:40 crc kubenswrapper[4799]: I1124 07:41:40.358837 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:40 crc kubenswrapper[4799]: I1124 07:41:40.431285 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ml8"] Nov 24 07:41:42 crc kubenswrapper[4799]: I1124 07:41:42.302169 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h4ml8" podUID="dcdca964-4f09-4026-926f-0abf69251368" containerName="registry-server" containerID="cri-o://e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712" gracePeriod=2 Nov 24 07:41:42 crc kubenswrapper[4799]: I1124 07:41:42.794337 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:42 crc kubenswrapper[4799]: I1124 07:41:42.925395 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-catalog-content\") pod \"dcdca964-4f09-4026-926f-0abf69251368\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " Nov 24 07:41:42 crc kubenswrapper[4799]: I1124 07:41:42.925454 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-utilities\") pod \"dcdca964-4f09-4026-926f-0abf69251368\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " Nov 24 07:41:42 crc kubenswrapper[4799]: I1124 07:41:42.925479 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4m4w\" (UniqueName: \"kubernetes.io/projected/dcdca964-4f09-4026-926f-0abf69251368-kube-api-access-f4m4w\") pod \"dcdca964-4f09-4026-926f-0abf69251368\" (UID: \"dcdca964-4f09-4026-926f-0abf69251368\") " Nov 24 07:41:42 crc kubenswrapper[4799]: I1124 07:41:42.927517 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-utilities" (OuterVolumeSpecName: "utilities") pod "dcdca964-4f09-4026-926f-0abf69251368" (UID: "dcdca964-4f09-4026-926f-0abf69251368"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:41:42 crc kubenswrapper[4799]: I1124 07:41:42.936091 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcdca964-4f09-4026-926f-0abf69251368-kube-api-access-f4m4w" (OuterVolumeSpecName: "kube-api-access-f4m4w") pod "dcdca964-4f09-4026-926f-0abf69251368" (UID: "dcdca964-4f09-4026-926f-0abf69251368"). InnerVolumeSpecName "kube-api-access-f4m4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:41:42 crc kubenswrapper[4799]: I1124 07:41:42.963125 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dcdca964-4f09-4026-926f-0abf69251368" (UID: "dcdca964-4f09-4026-926f-0abf69251368"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.027747 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.027884 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcdca964-4f09-4026-926f-0abf69251368-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.027932 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4m4w\" (UniqueName: \"kubernetes.io/projected/dcdca964-4f09-4026-926f-0abf69251368-kube-api-access-f4m4w\") on node \"crc\" DevicePath \"\"" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.317457 4799 generic.go:334] "Generic (PLEG): container finished" podID="dcdca964-4f09-4026-926f-0abf69251368" containerID="e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712" exitCode=0 Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.317527 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ml8" event={"ID":"dcdca964-4f09-4026-926f-0abf69251368","Type":"ContainerDied","Data":"e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712"} Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.317605 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ml8" event={"ID":"dcdca964-4f09-4026-926f-0abf69251368","Type":"ContainerDied","Data":"c4b836ef487e70dd404f9a38bcb11fda7425ab92374ea7c9ee711bad4f7d43e6"} Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.317602 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ml8" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.317655 4799 scope.go:117] "RemoveContainer" containerID="e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.348108 4799 scope.go:117] "RemoveContainer" containerID="87098c8ccfc2041a8fb43759cde2c2d463d48e0509784ddba5a2b4b774385621" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.386132 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ml8"] Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.389431 4799 scope.go:117] "RemoveContainer" containerID="8556a5ae88ceab5092ed484e0b2ac079b80b92ad817c9d304102039eecc19b4d" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.392771 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ml8"] Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.423811 4799 scope.go:117] "RemoveContainer" containerID="e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712" Nov 24 07:41:43 crc kubenswrapper[4799]: E1124 07:41:43.424351 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712\": container with ID starting with e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712 not found: ID does not exist" containerID="e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.424426 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712"} err="failed to get container status \"e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712\": rpc error: code = NotFound desc = could not find container \"e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712\": container with ID starting with e9a3b867e190ec8b73537d61b0e15f25e0f6729a39dff2a46f23e52982879712 not found: ID does not exist" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.424471 4799 scope.go:117] "RemoveContainer" containerID="87098c8ccfc2041a8fb43759cde2c2d463d48e0509784ddba5a2b4b774385621" Nov 24 07:41:43 crc kubenswrapper[4799]: E1124 07:41:43.424939 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87098c8ccfc2041a8fb43759cde2c2d463d48e0509784ddba5a2b4b774385621\": container with ID starting with 87098c8ccfc2041a8fb43759cde2c2d463d48e0509784ddba5a2b4b774385621 not found: ID does not exist" containerID="87098c8ccfc2041a8fb43759cde2c2d463d48e0509784ddba5a2b4b774385621" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.425107 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87098c8ccfc2041a8fb43759cde2c2d463d48e0509784ddba5a2b4b774385621"} err="failed to get container status \"87098c8ccfc2041a8fb43759cde2c2d463d48e0509784ddba5a2b4b774385621\": rpc error: code = NotFound desc = could not find container \"87098c8ccfc2041a8fb43759cde2c2d463d48e0509784ddba5a2b4b774385621\": container with ID starting with 87098c8ccfc2041a8fb43759cde2c2d463d48e0509784ddba5a2b4b774385621 not found: ID does not exist" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.425249 4799 scope.go:117] "RemoveContainer" containerID="8556a5ae88ceab5092ed484e0b2ac079b80b92ad817c9d304102039eecc19b4d" Nov 24 07:41:43 crc kubenswrapper[4799]: E1124 07:41:43.425736 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8556a5ae88ceab5092ed484e0b2ac079b80b92ad817c9d304102039eecc19b4d\": container with ID starting with 8556a5ae88ceab5092ed484e0b2ac079b80b92ad817c9d304102039eecc19b4d not found: ID does not exist" containerID="8556a5ae88ceab5092ed484e0b2ac079b80b92ad817c9d304102039eecc19b4d" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.425774 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8556a5ae88ceab5092ed484e0b2ac079b80b92ad817c9d304102039eecc19b4d"} err="failed to get container status \"8556a5ae88ceab5092ed484e0b2ac079b80b92ad817c9d304102039eecc19b4d\": rpc error: code = NotFound desc = could not find container \"8556a5ae88ceab5092ed484e0b2ac079b80b92ad817c9d304102039eecc19b4d\": container with ID starting with 8556a5ae88ceab5092ed484e0b2ac079b80b92ad817c9d304102039eecc19b4d not found: ID does not exist" Nov 24 07:41:43 crc kubenswrapper[4799]: I1124 07:41:43.643040 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcdca964-4f09-4026-926f-0abf69251368" path="/var/lib/kubelet/pods/dcdca964-4f09-4026-926f-0abf69251368/volumes" Nov 24 07:43:20 crc kubenswrapper[4799]: I1124 07:43:20.400257 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:43:20 crc kubenswrapper[4799]: I1124 07:43:20.401124 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.496908 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g9ptg"] Nov 24 07:43:30 crc kubenswrapper[4799]: E1124 07:43:30.498253 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcdca964-4f09-4026-926f-0abf69251368" containerName="extract-utilities" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.498282 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcdca964-4f09-4026-926f-0abf69251368" containerName="extract-utilities" Nov 24 07:43:30 crc kubenswrapper[4799]: E1124 07:43:30.498332 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcdca964-4f09-4026-926f-0abf69251368" containerName="extract-content" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.498374 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcdca964-4f09-4026-926f-0abf69251368" containerName="extract-content" Nov 24 07:43:30 crc kubenswrapper[4799]: E1124 07:43:30.498409 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcdca964-4f09-4026-926f-0abf69251368" containerName="registry-server" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.498421 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcdca964-4f09-4026-926f-0abf69251368" containerName="registry-server" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.498679 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcdca964-4f09-4026-926f-0abf69251368" containerName="registry-server" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.500492 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.520679 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g9ptg"] Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.642795 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-utilities\") pod \"community-operators-g9ptg\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.643260 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-catalog-content\") pod \"community-operators-g9ptg\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.643442 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwmth\" (UniqueName: \"kubernetes.io/projected/012b8160-69d7-4601-9aea-97e54518ecd2-kube-api-access-jwmth\") pod \"community-operators-g9ptg\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.744604 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-catalog-content\") pod \"community-operators-g9ptg\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.744910 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwmth\" (UniqueName: \"kubernetes.io/projected/012b8160-69d7-4601-9aea-97e54518ecd2-kube-api-access-jwmth\") pod \"community-operators-g9ptg\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.745092 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-utilities\") pod \"community-operators-g9ptg\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.745237 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-catalog-content\") pod \"community-operators-g9ptg\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.745621 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-utilities\") pod \"community-operators-g9ptg\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.768830 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwmth\" (UniqueName: \"kubernetes.io/projected/012b8160-69d7-4601-9aea-97e54518ecd2-kube-api-access-jwmth\") pod \"community-operators-g9ptg\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:30 crc kubenswrapper[4799]: I1124 07:43:30.842120 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:31 crc kubenswrapper[4799]: I1124 07:43:31.164683 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g9ptg"] Nov 24 07:43:31 crc kubenswrapper[4799]: I1124 07:43:31.557813 4799 generic.go:334] "Generic (PLEG): container finished" podID="012b8160-69d7-4601-9aea-97e54518ecd2" containerID="2f9bf0cd6757217cadfb551dc2cdc6de27c6dbd0ffea3fa60e3214ee49aeee0e" exitCode=0 Nov 24 07:43:31 crc kubenswrapper[4799]: I1124 07:43:31.557958 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9ptg" event={"ID":"012b8160-69d7-4601-9aea-97e54518ecd2","Type":"ContainerDied","Data":"2f9bf0cd6757217cadfb551dc2cdc6de27c6dbd0ffea3fa60e3214ee49aeee0e"} Nov 24 07:43:31 crc kubenswrapper[4799]: I1124 07:43:31.557984 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9ptg" event={"ID":"012b8160-69d7-4601-9aea-97e54518ecd2","Type":"ContainerStarted","Data":"5ad5573ce98287b9e2250fc7224bc85a5be8e47868643de653898912ead3ffb1"} Nov 24 07:43:31 crc kubenswrapper[4799]: I1124 07:43:31.559927 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:43:32 crc kubenswrapper[4799]: I1124 07:43:32.568018 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9ptg" event={"ID":"012b8160-69d7-4601-9aea-97e54518ecd2","Type":"ContainerStarted","Data":"f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51"} Nov 24 07:43:33 crc kubenswrapper[4799]: I1124 07:43:33.579557 4799 generic.go:334] "Generic (PLEG): container finished" podID="012b8160-69d7-4601-9aea-97e54518ecd2" containerID="f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51" exitCode=0 Nov 24 07:43:33 crc kubenswrapper[4799]: I1124 07:43:33.579610 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9ptg" event={"ID":"012b8160-69d7-4601-9aea-97e54518ecd2","Type":"ContainerDied","Data":"f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51"} Nov 24 07:43:34 crc kubenswrapper[4799]: I1124 07:43:34.597296 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9ptg" event={"ID":"012b8160-69d7-4601-9aea-97e54518ecd2","Type":"ContainerStarted","Data":"5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82"} Nov 24 07:43:34 crc kubenswrapper[4799]: I1124 07:43:34.621190 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g9ptg" podStartSLOduration=2.168063563 podStartE2EDuration="4.621166597s" podCreationTimestamp="2025-11-24 07:43:30 +0000 UTC" firstStartedPulling="2025-11-24 07:43:31.559681658 +0000 UTC m=+3357.215664132" lastFinishedPulling="2025-11-24 07:43:34.012784672 +0000 UTC m=+3359.668767166" observedRunningTime="2025-11-24 07:43:34.614979762 +0000 UTC m=+3360.270962256" watchObservedRunningTime="2025-11-24 07:43:34.621166597 +0000 UTC m=+3360.277149081" Nov 24 07:43:40 crc kubenswrapper[4799]: I1124 07:43:40.843118 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:40 crc kubenswrapper[4799]: I1124 07:43:40.846090 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:40 crc kubenswrapper[4799]: I1124 07:43:40.905683 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:41 crc kubenswrapper[4799]: I1124 07:43:41.738311 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:41 crc kubenswrapper[4799]: I1124 07:43:41.805981 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g9ptg"] Nov 24 07:43:43 crc kubenswrapper[4799]: I1124 07:43:43.683013 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g9ptg" podUID="012b8160-69d7-4601-9aea-97e54518ecd2" containerName="registry-server" containerID="cri-o://5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82" gracePeriod=2 Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.130525 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.238225 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwmth\" (UniqueName: \"kubernetes.io/projected/012b8160-69d7-4601-9aea-97e54518ecd2-kube-api-access-jwmth\") pod \"012b8160-69d7-4601-9aea-97e54518ecd2\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.238338 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-utilities\") pod \"012b8160-69d7-4601-9aea-97e54518ecd2\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.238361 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-catalog-content\") pod \"012b8160-69d7-4601-9aea-97e54518ecd2\" (UID: \"012b8160-69d7-4601-9aea-97e54518ecd2\") " Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.239321 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-utilities" (OuterVolumeSpecName: "utilities") pod "012b8160-69d7-4601-9aea-97e54518ecd2" (UID: "012b8160-69d7-4601-9aea-97e54518ecd2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.244972 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/012b8160-69d7-4601-9aea-97e54518ecd2-kube-api-access-jwmth" (OuterVolumeSpecName: "kube-api-access-jwmth") pod "012b8160-69d7-4601-9aea-97e54518ecd2" (UID: "012b8160-69d7-4601-9aea-97e54518ecd2"). InnerVolumeSpecName "kube-api-access-jwmth". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.287234 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "012b8160-69d7-4601-9aea-97e54518ecd2" (UID: "012b8160-69d7-4601-9aea-97e54518ecd2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.340399 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwmth\" (UniqueName: \"kubernetes.io/projected/012b8160-69d7-4601-9aea-97e54518ecd2-kube-api-access-jwmth\") on node \"crc\" DevicePath \"\"" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.340436 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.340448 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/012b8160-69d7-4601-9aea-97e54518ecd2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.693392 4799 generic.go:334] "Generic (PLEG): container finished" podID="012b8160-69d7-4601-9aea-97e54518ecd2" containerID="5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82" exitCode=0 Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.693462 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9ptg" event={"ID":"012b8160-69d7-4601-9aea-97e54518ecd2","Type":"ContainerDied","Data":"5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82"} Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.693505 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9ptg" event={"ID":"012b8160-69d7-4601-9aea-97e54518ecd2","Type":"ContainerDied","Data":"5ad5573ce98287b9e2250fc7224bc85a5be8e47868643de653898912ead3ffb1"} Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.693534 4799 scope.go:117] "RemoveContainer" containerID="5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.693733 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9ptg" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.710395 4799 scope.go:117] "RemoveContainer" containerID="f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.751596 4799 scope.go:117] "RemoveContainer" containerID="2f9bf0cd6757217cadfb551dc2cdc6de27c6dbd0ffea3fa60e3214ee49aeee0e" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.758303 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g9ptg"] Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.767373 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g9ptg"] Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.770205 4799 scope.go:117] "RemoveContainer" containerID="5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82" Nov 24 07:43:44 crc kubenswrapper[4799]: E1124 07:43:44.770664 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82\": container with ID starting with 5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82 not found: ID does not exist" containerID="5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.770701 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82"} err="failed to get container status \"5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82\": rpc error: code = NotFound desc = could not find container \"5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82\": container with ID starting with 5a9b4385483e25257c00e26c47ff4e8d1d4d0dfc0876cec2853d381c985b1c82 not found: ID does not exist" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.770737 4799 scope.go:117] "RemoveContainer" containerID="f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51" Nov 24 07:43:44 crc kubenswrapper[4799]: E1124 07:43:44.770998 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51\": container with ID starting with f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51 not found: ID does not exist" containerID="f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.771045 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51"} err="failed to get container status \"f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51\": rpc error: code = NotFound desc = could not find container \"f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51\": container with ID starting with f57b562cfdf38981ea7fe895f27f334b02b3e9a28b665c9f202c38b56c8f9a51 not found: ID does not exist" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.771077 4799 scope.go:117] "RemoveContainer" containerID="2f9bf0cd6757217cadfb551dc2cdc6de27c6dbd0ffea3fa60e3214ee49aeee0e" Nov 24 07:43:44 crc kubenswrapper[4799]: E1124 07:43:44.771385 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f9bf0cd6757217cadfb551dc2cdc6de27c6dbd0ffea3fa60e3214ee49aeee0e\": container with ID starting with 2f9bf0cd6757217cadfb551dc2cdc6de27c6dbd0ffea3fa60e3214ee49aeee0e not found: ID does not exist" containerID="2f9bf0cd6757217cadfb551dc2cdc6de27c6dbd0ffea3fa60e3214ee49aeee0e" Nov 24 07:43:44 crc kubenswrapper[4799]: I1124 07:43:44.771410 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f9bf0cd6757217cadfb551dc2cdc6de27c6dbd0ffea3fa60e3214ee49aeee0e"} err="failed to get container status \"2f9bf0cd6757217cadfb551dc2cdc6de27c6dbd0ffea3fa60e3214ee49aeee0e\": rpc error: code = NotFound desc = could not find container \"2f9bf0cd6757217cadfb551dc2cdc6de27c6dbd0ffea3fa60e3214ee49aeee0e\": container with ID starting with 2f9bf0cd6757217cadfb551dc2cdc6de27c6dbd0ffea3fa60e3214ee49aeee0e not found: ID does not exist" Nov 24 07:43:45 crc kubenswrapper[4799]: I1124 07:43:45.645495 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="012b8160-69d7-4601-9aea-97e54518ecd2" path="/var/lib/kubelet/pods/012b8160-69d7-4601-9aea-97e54518ecd2/volumes" Nov 24 07:43:50 crc kubenswrapper[4799]: I1124 07:43:50.401380 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:43:50 crc kubenswrapper[4799]: I1124 07:43:50.401738 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:44:20 crc kubenswrapper[4799]: I1124 07:44:20.400163 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:44:20 crc kubenswrapper[4799]: I1124 07:44:20.400732 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:44:20 crc kubenswrapper[4799]: I1124 07:44:20.400787 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:44:20 crc kubenswrapper[4799]: I1124 07:44:20.401533 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"80a5fa14424f6f54364c175b542535d6d16c363682f8b22f6fe58a3c63c8ac33"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:44:20 crc kubenswrapper[4799]: I1124 07:44:20.401602 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://80a5fa14424f6f54364c175b542535d6d16c363682f8b22f6fe58a3c63c8ac33" gracePeriod=600 Nov 24 07:44:21 crc kubenswrapper[4799]: I1124 07:44:21.022033 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="80a5fa14424f6f54364c175b542535d6d16c363682f8b22f6fe58a3c63c8ac33" exitCode=0 Nov 24 07:44:21 crc kubenswrapper[4799]: I1124 07:44:21.022114 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"80a5fa14424f6f54364c175b542535d6d16c363682f8b22f6fe58a3c63c8ac33"} Nov 24 07:44:21 crc kubenswrapper[4799]: I1124 07:44:21.022377 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832"} Nov 24 07:44:21 crc kubenswrapper[4799]: I1124 07:44:21.022416 4799 scope.go:117] "RemoveContainer" containerID="5fcb624d314b8b362598ca3da56fd8984c5db38758cb8eadb0cde38a82c55882" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.210564 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb"] Nov 24 07:45:00 crc kubenswrapper[4799]: E1124 07:45:00.211644 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="012b8160-69d7-4601-9aea-97e54518ecd2" containerName="extract-utilities" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.211661 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="012b8160-69d7-4601-9aea-97e54518ecd2" containerName="extract-utilities" Nov 24 07:45:00 crc kubenswrapper[4799]: E1124 07:45:00.211680 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="012b8160-69d7-4601-9aea-97e54518ecd2" containerName="registry-server" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.211689 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="012b8160-69d7-4601-9aea-97e54518ecd2" containerName="registry-server" Nov 24 07:45:00 crc kubenswrapper[4799]: E1124 07:45:00.211708 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="012b8160-69d7-4601-9aea-97e54518ecd2" containerName="extract-content" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.211717 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="012b8160-69d7-4601-9aea-97e54518ecd2" containerName="extract-content" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.211946 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="012b8160-69d7-4601-9aea-97e54518ecd2" containerName="registry-server" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.212547 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.215644 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.215925 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.216418 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb"] Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.407014 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2772baa3-685c-40ae-b995-757ee71a5fcd-config-volume\") pod \"collect-profiles-29399505-b2gvb\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.407085 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2772baa3-685c-40ae-b995-757ee71a5fcd-secret-volume\") pod \"collect-profiles-29399505-b2gvb\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.407221 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s9z2\" (UniqueName: \"kubernetes.io/projected/2772baa3-685c-40ae-b995-757ee71a5fcd-kube-api-access-9s9z2\") pod \"collect-profiles-29399505-b2gvb\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.509280 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2772baa3-685c-40ae-b995-757ee71a5fcd-config-volume\") pod \"collect-profiles-29399505-b2gvb\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.509400 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2772baa3-685c-40ae-b995-757ee71a5fcd-secret-volume\") pod \"collect-profiles-29399505-b2gvb\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.509631 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s9z2\" (UniqueName: \"kubernetes.io/projected/2772baa3-685c-40ae-b995-757ee71a5fcd-kube-api-access-9s9z2\") pod \"collect-profiles-29399505-b2gvb\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.511207 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2772baa3-685c-40ae-b995-757ee71a5fcd-config-volume\") pod \"collect-profiles-29399505-b2gvb\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.518645 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2772baa3-685c-40ae-b995-757ee71a5fcd-secret-volume\") pod \"collect-profiles-29399505-b2gvb\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.527707 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s9z2\" (UniqueName: \"kubernetes.io/projected/2772baa3-685c-40ae-b995-757ee71a5fcd-kube-api-access-9s9z2\") pod \"collect-profiles-29399505-b2gvb\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.544676 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:00 crc kubenswrapper[4799]: I1124 07:45:00.994109 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb"] Nov 24 07:45:00 crc kubenswrapper[4799]: W1124 07:45:00.996891 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2772baa3_685c_40ae_b995_757ee71a5fcd.slice/crio-0ef2a20b53df5ac42cf23151696f4c609d794763a8c3f665a6a34e69a2ac8032 WatchSource:0}: Error finding container 0ef2a20b53df5ac42cf23151696f4c609d794763a8c3f665a6a34e69a2ac8032: Status 404 returned error can't find the container with id 0ef2a20b53df5ac42cf23151696f4c609d794763a8c3f665a6a34e69a2ac8032 Nov 24 07:45:01 crc kubenswrapper[4799]: I1124 07:45:01.412046 4799 generic.go:334] "Generic (PLEG): container finished" podID="2772baa3-685c-40ae-b995-757ee71a5fcd" containerID="7f870487cd6f8acf504f8baca030981897011639e9ddd4b22dd7dcf6e30952d5" exitCode=0 Nov 24 07:45:01 crc kubenswrapper[4799]: I1124 07:45:01.412330 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" event={"ID":"2772baa3-685c-40ae-b995-757ee71a5fcd","Type":"ContainerDied","Data":"7f870487cd6f8acf504f8baca030981897011639e9ddd4b22dd7dcf6e30952d5"} Nov 24 07:45:01 crc kubenswrapper[4799]: I1124 07:45:01.412353 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" event={"ID":"2772baa3-685c-40ae-b995-757ee71a5fcd","Type":"ContainerStarted","Data":"0ef2a20b53df5ac42cf23151696f4c609d794763a8c3f665a6a34e69a2ac8032"} Nov 24 07:45:02 crc kubenswrapper[4799]: I1124 07:45:02.808535 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:02 crc kubenswrapper[4799]: I1124 07:45:02.945756 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2772baa3-685c-40ae-b995-757ee71a5fcd-secret-volume\") pod \"2772baa3-685c-40ae-b995-757ee71a5fcd\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " Nov 24 07:45:02 crc kubenswrapper[4799]: I1124 07:45:02.945925 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s9z2\" (UniqueName: \"kubernetes.io/projected/2772baa3-685c-40ae-b995-757ee71a5fcd-kube-api-access-9s9z2\") pod \"2772baa3-685c-40ae-b995-757ee71a5fcd\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " Nov 24 07:45:02 crc kubenswrapper[4799]: I1124 07:45:02.945946 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2772baa3-685c-40ae-b995-757ee71a5fcd-config-volume\") pod \"2772baa3-685c-40ae-b995-757ee71a5fcd\" (UID: \"2772baa3-685c-40ae-b995-757ee71a5fcd\") " Nov 24 07:45:02 crc kubenswrapper[4799]: I1124 07:45:02.946486 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2772baa3-685c-40ae-b995-757ee71a5fcd-config-volume" (OuterVolumeSpecName: "config-volume") pod "2772baa3-685c-40ae-b995-757ee71a5fcd" (UID: "2772baa3-685c-40ae-b995-757ee71a5fcd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:45:02 crc kubenswrapper[4799]: I1124 07:45:02.952145 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2772baa3-685c-40ae-b995-757ee71a5fcd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2772baa3-685c-40ae-b995-757ee71a5fcd" (UID: "2772baa3-685c-40ae-b995-757ee71a5fcd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:45:02 crc kubenswrapper[4799]: I1124 07:45:02.953586 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2772baa3-685c-40ae-b995-757ee71a5fcd-kube-api-access-9s9z2" (OuterVolumeSpecName: "kube-api-access-9s9z2") pod "2772baa3-685c-40ae-b995-757ee71a5fcd" (UID: "2772baa3-685c-40ae-b995-757ee71a5fcd"). InnerVolumeSpecName "kube-api-access-9s9z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:45:03 crc kubenswrapper[4799]: I1124 07:45:03.048191 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s9z2\" (UniqueName: \"kubernetes.io/projected/2772baa3-685c-40ae-b995-757ee71a5fcd-kube-api-access-9s9z2\") on node \"crc\" DevicePath \"\"" Nov 24 07:45:03 crc kubenswrapper[4799]: I1124 07:45:03.048252 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2772baa3-685c-40ae-b995-757ee71a5fcd-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:45:03 crc kubenswrapper[4799]: I1124 07:45:03.048278 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2772baa3-685c-40ae-b995-757ee71a5fcd-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 07:45:03 crc kubenswrapper[4799]: I1124 07:45:03.431545 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" event={"ID":"2772baa3-685c-40ae-b995-757ee71a5fcd","Type":"ContainerDied","Data":"0ef2a20b53df5ac42cf23151696f4c609d794763a8c3f665a6a34e69a2ac8032"} Nov 24 07:45:03 crc kubenswrapper[4799]: I1124 07:45:03.431608 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ef2a20b53df5ac42cf23151696f4c609d794763a8c3f665a6a34e69a2ac8032" Nov 24 07:45:03 crc kubenswrapper[4799]: I1124 07:45:03.431635 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb" Nov 24 07:45:03 crc kubenswrapper[4799]: I1124 07:45:03.922162 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5"] Nov 24 07:45:03 crc kubenswrapper[4799]: I1124 07:45:03.930566 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399460-tqpm5"] Nov 24 07:45:05 crc kubenswrapper[4799]: I1124 07:45:05.648284 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0570ad32-6b7f-44f3-90aa-1cac51750548" path="/var/lib/kubelet/pods/0570ad32-6b7f-44f3-90aa-1cac51750548/volumes" Nov 24 07:45:10 crc kubenswrapper[4799]: I1124 07:45:10.870444 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lpj46"] Nov 24 07:45:10 crc kubenswrapper[4799]: E1124 07:45:10.872090 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2772baa3-685c-40ae-b995-757ee71a5fcd" containerName="collect-profiles" Nov 24 07:45:10 crc kubenswrapper[4799]: I1124 07:45:10.872125 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2772baa3-685c-40ae-b995-757ee71a5fcd" containerName="collect-profiles" Nov 24 07:45:10 crc kubenswrapper[4799]: I1124 07:45:10.872469 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2772baa3-685c-40ae-b995-757ee71a5fcd" containerName="collect-profiles" Nov 24 07:45:10 crc kubenswrapper[4799]: I1124 07:45:10.874998 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:10 crc kubenswrapper[4799]: I1124 07:45:10.884668 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lpj46"] Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.023812 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpxh2\" (UniqueName: \"kubernetes.io/projected/f0a396a7-ff05-4817-b53f-ba66c85eae6e-kube-api-access-lpxh2\") pod \"redhat-operators-lpj46\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.023959 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-catalog-content\") pod \"redhat-operators-lpj46\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.024024 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-utilities\") pod \"redhat-operators-lpj46\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.125389 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpxh2\" (UniqueName: \"kubernetes.io/projected/f0a396a7-ff05-4817-b53f-ba66c85eae6e-kube-api-access-lpxh2\") pod \"redhat-operators-lpj46\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.125475 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-catalog-content\") pod \"redhat-operators-lpj46\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.125518 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-utilities\") pod \"redhat-operators-lpj46\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.126186 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-utilities\") pod \"redhat-operators-lpj46\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.126522 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-catalog-content\") pod \"redhat-operators-lpj46\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.151669 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpxh2\" (UniqueName: \"kubernetes.io/projected/f0a396a7-ff05-4817-b53f-ba66c85eae6e-kube-api-access-lpxh2\") pod \"redhat-operators-lpj46\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.232001 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.485119 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lpj46"] Nov 24 07:45:11 crc kubenswrapper[4799]: I1124 07:45:11.503605 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpj46" event={"ID":"f0a396a7-ff05-4817-b53f-ba66c85eae6e","Type":"ContainerStarted","Data":"7467ca7140e8cd5e4bb4dc094a43d5184071abc77df90fbde2c1a3a9233a9c38"} Nov 24 07:45:12 crc kubenswrapper[4799]: I1124 07:45:12.525264 4799 generic.go:334] "Generic (PLEG): container finished" podID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerID="65c847b772ac64797391f43377db76bd3057b86f5c3d21b34e2b4d4ba8cb84ad" exitCode=0 Nov 24 07:45:12 crc kubenswrapper[4799]: I1124 07:45:12.525413 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpj46" event={"ID":"f0a396a7-ff05-4817-b53f-ba66c85eae6e","Type":"ContainerDied","Data":"65c847b772ac64797391f43377db76bd3057b86f5c3d21b34e2b4d4ba8cb84ad"} Nov 24 07:45:13 crc kubenswrapper[4799]: I1124 07:45:13.546767 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpj46" event={"ID":"f0a396a7-ff05-4817-b53f-ba66c85eae6e","Type":"ContainerStarted","Data":"a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f"} Nov 24 07:45:14 crc kubenswrapper[4799]: I1124 07:45:14.559286 4799 generic.go:334] "Generic (PLEG): container finished" podID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerID="a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f" exitCode=0 Nov 24 07:45:14 crc kubenswrapper[4799]: I1124 07:45:14.559357 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpj46" event={"ID":"f0a396a7-ff05-4817-b53f-ba66c85eae6e","Type":"ContainerDied","Data":"a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f"} Nov 24 07:45:15 crc kubenswrapper[4799]: I1124 07:45:15.572408 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpj46" event={"ID":"f0a396a7-ff05-4817-b53f-ba66c85eae6e","Type":"ContainerStarted","Data":"7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e"} Nov 24 07:45:15 crc kubenswrapper[4799]: I1124 07:45:15.596319 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lpj46" podStartSLOduration=3.035923164 podStartE2EDuration="5.596297502s" podCreationTimestamp="2025-11-24 07:45:10 +0000 UTC" firstStartedPulling="2025-11-24 07:45:12.527460637 +0000 UTC m=+3458.183443151" lastFinishedPulling="2025-11-24 07:45:15.087834995 +0000 UTC m=+3460.743817489" observedRunningTime="2025-11-24 07:45:15.592780692 +0000 UTC m=+3461.248763176" watchObservedRunningTime="2025-11-24 07:45:15.596297502 +0000 UTC m=+3461.252279996" Nov 24 07:45:21 crc kubenswrapper[4799]: I1124 07:45:21.233073 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:21 crc kubenswrapper[4799]: I1124 07:45:21.233790 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:22 crc kubenswrapper[4799]: I1124 07:45:22.305350 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lpj46" podUID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerName="registry-server" probeResult="failure" output=< Nov 24 07:45:22 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 07:45:22 crc kubenswrapper[4799]: > Nov 24 07:45:31 crc kubenswrapper[4799]: I1124 07:45:31.287540 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:31 crc kubenswrapper[4799]: I1124 07:45:31.334035 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:31 crc kubenswrapper[4799]: I1124 07:45:31.524780 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lpj46"] Nov 24 07:45:32 crc kubenswrapper[4799]: I1124 07:45:32.722703 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lpj46" podUID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerName="registry-server" containerID="cri-o://7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e" gracePeriod=2 Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.182433 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.317328 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-catalog-content\") pod \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.317578 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpxh2\" (UniqueName: \"kubernetes.io/projected/f0a396a7-ff05-4817-b53f-ba66c85eae6e-kube-api-access-lpxh2\") pod \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.317688 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-utilities\") pod \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\" (UID: \"f0a396a7-ff05-4817-b53f-ba66c85eae6e\") " Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.318557 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-utilities" (OuterVolumeSpecName: "utilities") pod "f0a396a7-ff05-4817-b53f-ba66c85eae6e" (UID: "f0a396a7-ff05-4817-b53f-ba66c85eae6e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.325322 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0a396a7-ff05-4817-b53f-ba66c85eae6e-kube-api-access-lpxh2" (OuterVolumeSpecName: "kube-api-access-lpxh2") pod "f0a396a7-ff05-4817-b53f-ba66c85eae6e" (UID: "f0a396a7-ff05-4817-b53f-ba66c85eae6e"). InnerVolumeSpecName "kube-api-access-lpxh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.419968 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpxh2\" (UniqueName: \"kubernetes.io/projected/f0a396a7-ff05-4817-b53f-ba66c85eae6e-kube-api-access-lpxh2\") on node \"crc\" DevicePath \"\"" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.420366 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.445697 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0a396a7-ff05-4817-b53f-ba66c85eae6e" (UID: "f0a396a7-ff05-4817-b53f-ba66c85eae6e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.522722 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a396a7-ff05-4817-b53f-ba66c85eae6e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.731696 4799 generic.go:334] "Generic (PLEG): container finished" podID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerID="7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e" exitCode=0 Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.731738 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpj46" event={"ID":"f0a396a7-ff05-4817-b53f-ba66c85eae6e","Type":"ContainerDied","Data":"7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e"} Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.731767 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lpj46" event={"ID":"f0a396a7-ff05-4817-b53f-ba66c85eae6e","Type":"ContainerDied","Data":"7467ca7140e8cd5e4bb4dc094a43d5184071abc77df90fbde2c1a3a9233a9c38"} Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.731785 4799 scope.go:117] "RemoveContainer" containerID="7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.731940 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lpj46" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.754957 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lpj46"] Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.758362 4799 scope.go:117] "RemoveContainer" containerID="a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.760313 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lpj46"] Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.782686 4799 scope.go:117] "RemoveContainer" containerID="65c847b772ac64797391f43377db76bd3057b86f5c3d21b34e2b4d4ba8cb84ad" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.810888 4799 scope.go:117] "RemoveContainer" containerID="7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e" Nov 24 07:45:33 crc kubenswrapper[4799]: E1124 07:45:33.811327 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e\": container with ID starting with 7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e not found: ID does not exist" containerID="7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.811362 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e"} err="failed to get container status \"7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e\": rpc error: code = NotFound desc = could not find container \"7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e\": container with ID starting with 7bd749f80f1f7e18ca79a01a15a50de462acf40141b74b1cc2674e17e6a0c40e not found: ID does not exist" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.811385 4799 scope.go:117] "RemoveContainer" containerID="a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f" Nov 24 07:45:33 crc kubenswrapper[4799]: E1124 07:45:33.812680 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f\": container with ID starting with a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f not found: ID does not exist" containerID="a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.812715 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f"} err="failed to get container status \"a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f\": rpc error: code = NotFound desc = could not find container \"a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f\": container with ID starting with a1ea09c362be3a3b35b8e2b6bad0e130dd3f3cab3f5baab7d86b65784140fc1f not found: ID does not exist" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.812734 4799 scope.go:117] "RemoveContainer" containerID="65c847b772ac64797391f43377db76bd3057b86f5c3d21b34e2b4d4ba8cb84ad" Nov 24 07:45:33 crc kubenswrapper[4799]: E1124 07:45:33.813037 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65c847b772ac64797391f43377db76bd3057b86f5c3d21b34e2b4d4ba8cb84ad\": container with ID starting with 65c847b772ac64797391f43377db76bd3057b86f5c3d21b34e2b4d4ba8cb84ad not found: ID does not exist" containerID="65c847b772ac64797391f43377db76bd3057b86f5c3d21b34e2b4d4ba8cb84ad" Nov 24 07:45:33 crc kubenswrapper[4799]: I1124 07:45:33.813064 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65c847b772ac64797391f43377db76bd3057b86f5c3d21b34e2b4d4ba8cb84ad"} err="failed to get container status \"65c847b772ac64797391f43377db76bd3057b86f5c3d21b34e2b4d4ba8cb84ad\": rpc error: code = NotFound desc = could not find container \"65c847b772ac64797391f43377db76bd3057b86f5c3d21b34e2b4d4ba8cb84ad\": container with ID starting with 65c847b772ac64797391f43377db76bd3057b86f5c3d21b34e2b4d4ba8cb84ad not found: ID does not exist" Nov 24 07:45:35 crc kubenswrapper[4799]: I1124 07:45:35.645169 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" path="/var/lib/kubelet/pods/f0a396a7-ff05-4817-b53f-ba66c85eae6e/volumes" Nov 24 07:45:53 crc kubenswrapper[4799]: I1124 07:45:53.540237 4799 scope.go:117] "RemoveContainer" containerID="8682d1cf7ee42852f953abfeecee75629e7ea60dc5dc60582f4a0c0b778ff7fa" Nov 24 07:46:20 crc kubenswrapper[4799]: I1124 07:46:20.400917 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:46:20 crc kubenswrapper[4799]: I1124 07:46:20.401735 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:46:50 crc kubenswrapper[4799]: I1124 07:46:50.400334 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:46:50 crc kubenswrapper[4799]: I1124 07:46:50.401165 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.024621 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2cg96"] Nov 24 07:47:20 crc kubenswrapper[4799]: E1124 07:47:20.029099 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerName="extract-utilities" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.029141 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerName="extract-utilities" Nov 24 07:47:20 crc kubenswrapper[4799]: E1124 07:47:20.029177 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerName="registry-server" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.029198 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerName="registry-server" Nov 24 07:47:20 crc kubenswrapper[4799]: E1124 07:47:20.029271 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerName="extract-content" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.029292 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerName="extract-content" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.029914 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0a396a7-ff05-4817-b53f-ba66c85eae6e" containerName="registry-server" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.033691 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.037538 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2cg96"] Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.182013 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-catalog-content\") pod \"certified-operators-2cg96\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.182052 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-utilities\") pod \"certified-operators-2cg96\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.182088 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnmnx\" (UniqueName: \"kubernetes.io/projected/3f10cb59-0728-4c9d-b104-c71b65563950-kube-api-access-wnmnx\") pod \"certified-operators-2cg96\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.283136 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnmnx\" (UniqueName: \"kubernetes.io/projected/3f10cb59-0728-4c9d-b104-c71b65563950-kube-api-access-wnmnx\") pod \"certified-operators-2cg96\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.283278 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-catalog-content\") pod \"certified-operators-2cg96\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.283303 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-utilities\") pod \"certified-operators-2cg96\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.283785 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-utilities\") pod \"certified-operators-2cg96\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.284098 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-catalog-content\") pod \"certified-operators-2cg96\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.314919 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnmnx\" (UniqueName: \"kubernetes.io/projected/3f10cb59-0728-4c9d-b104-c71b65563950-kube-api-access-wnmnx\") pod \"certified-operators-2cg96\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.374483 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.400464 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.400521 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.400573 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.401199 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.401258 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" gracePeriod=600 Nov 24 07:47:20 crc kubenswrapper[4799]: E1124 07:47:20.569032 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.741735 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" exitCode=0 Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.741777 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832"} Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.741808 4799 scope.go:117] "RemoveContainer" containerID="80a5fa14424f6f54364c175b542535d6d16c363682f8b22f6fe58a3c63c8ac33" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.742237 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:47:20 crc kubenswrapper[4799]: E1124 07:47:20.742481 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:47:20 crc kubenswrapper[4799]: I1124 07:47:20.849930 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2cg96"] Nov 24 07:47:21 crc kubenswrapper[4799]: I1124 07:47:21.752162 4799 generic.go:334] "Generic (PLEG): container finished" podID="3f10cb59-0728-4c9d-b104-c71b65563950" containerID="86139a0afdb7e9b1522fc94ff46ebbc967827ec9eb2832f5df2718834704705d" exitCode=0 Nov 24 07:47:21 crc kubenswrapper[4799]: I1124 07:47:21.752254 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cg96" event={"ID":"3f10cb59-0728-4c9d-b104-c71b65563950","Type":"ContainerDied","Data":"86139a0afdb7e9b1522fc94ff46ebbc967827ec9eb2832f5df2718834704705d"} Nov 24 07:47:21 crc kubenswrapper[4799]: I1124 07:47:21.752768 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cg96" event={"ID":"3f10cb59-0728-4c9d-b104-c71b65563950","Type":"ContainerStarted","Data":"b4835b735b47df65df47bdda778cd012971ea188bdf0a98f153ab4811552bf57"} Nov 24 07:47:22 crc kubenswrapper[4799]: I1124 07:47:22.770342 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cg96" event={"ID":"3f10cb59-0728-4c9d-b104-c71b65563950","Type":"ContainerStarted","Data":"3cec9c21bcd7ccf8f36fc17ed0a663c1fa4b6692d475d6510b1a131a500d1c84"} Nov 24 07:47:23 crc kubenswrapper[4799]: I1124 07:47:23.783483 4799 generic.go:334] "Generic (PLEG): container finished" podID="3f10cb59-0728-4c9d-b104-c71b65563950" containerID="3cec9c21bcd7ccf8f36fc17ed0a663c1fa4b6692d475d6510b1a131a500d1c84" exitCode=0 Nov 24 07:47:23 crc kubenswrapper[4799]: I1124 07:47:23.783592 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cg96" event={"ID":"3f10cb59-0728-4c9d-b104-c71b65563950","Type":"ContainerDied","Data":"3cec9c21bcd7ccf8f36fc17ed0a663c1fa4b6692d475d6510b1a131a500d1c84"} Nov 24 07:47:24 crc kubenswrapper[4799]: I1124 07:47:24.799401 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cg96" event={"ID":"3f10cb59-0728-4c9d-b104-c71b65563950","Type":"ContainerStarted","Data":"eedc52f33720d3cf8d8151e4505ec6de3450715f191355ee06eaa8418217f794"} Nov 24 07:47:24 crc kubenswrapper[4799]: I1124 07:47:24.821420 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2cg96" podStartSLOduration=3.371452018 podStartE2EDuration="5.821389446s" podCreationTimestamp="2025-11-24 07:47:19 +0000 UTC" firstStartedPulling="2025-11-24 07:47:21.755436083 +0000 UTC m=+3587.411418597" lastFinishedPulling="2025-11-24 07:47:24.205373511 +0000 UTC m=+3589.861356025" observedRunningTime="2025-11-24 07:47:24.818917026 +0000 UTC m=+3590.474899500" watchObservedRunningTime="2025-11-24 07:47:24.821389446 +0000 UTC m=+3590.477371960" Nov 24 07:47:30 crc kubenswrapper[4799]: I1124 07:47:30.374829 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:30 crc kubenswrapper[4799]: I1124 07:47:30.377444 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:30 crc kubenswrapper[4799]: I1124 07:47:30.435766 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:30 crc kubenswrapper[4799]: I1124 07:47:30.944148 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:31 crc kubenswrapper[4799]: I1124 07:47:31.001449 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2cg96"] Nov 24 07:47:32 crc kubenswrapper[4799]: I1124 07:47:32.880220 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2cg96" podUID="3f10cb59-0728-4c9d-b104-c71b65563950" containerName="registry-server" containerID="cri-o://eedc52f33720d3cf8d8151e4505ec6de3450715f191355ee06eaa8418217f794" gracePeriod=2 Nov 24 07:47:33 crc kubenswrapper[4799]: I1124 07:47:33.892234 4799 generic.go:334] "Generic (PLEG): container finished" podID="3f10cb59-0728-4c9d-b104-c71b65563950" containerID="eedc52f33720d3cf8d8151e4505ec6de3450715f191355ee06eaa8418217f794" exitCode=0 Nov 24 07:47:33 crc kubenswrapper[4799]: I1124 07:47:33.892296 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cg96" event={"ID":"3f10cb59-0728-4c9d-b104-c71b65563950","Type":"ContainerDied","Data":"eedc52f33720d3cf8d8151e4505ec6de3450715f191355ee06eaa8418217f794"} Nov 24 07:47:33 crc kubenswrapper[4799]: I1124 07:47:33.892891 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cg96" event={"ID":"3f10cb59-0728-4c9d-b104-c71b65563950","Type":"ContainerDied","Data":"b4835b735b47df65df47bdda778cd012971ea188bdf0a98f153ab4811552bf57"} Nov 24 07:47:33 crc kubenswrapper[4799]: I1124 07:47:33.892918 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4835b735b47df65df47bdda778cd012971ea188bdf0a98f153ab4811552bf57" Nov 24 07:47:33 crc kubenswrapper[4799]: I1124 07:47:33.944540 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:33 crc kubenswrapper[4799]: I1124 07:47:33.998555 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-catalog-content\") pod \"3f10cb59-0728-4c9d-b104-c71b65563950\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " Nov 24 07:47:33 crc kubenswrapper[4799]: I1124 07:47:33.998614 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-utilities\") pod \"3f10cb59-0728-4c9d-b104-c71b65563950\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " Nov 24 07:47:33 crc kubenswrapper[4799]: I1124 07:47:33.998685 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnmnx\" (UniqueName: \"kubernetes.io/projected/3f10cb59-0728-4c9d-b104-c71b65563950-kube-api-access-wnmnx\") pod \"3f10cb59-0728-4c9d-b104-c71b65563950\" (UID: \"3f10cb59-0728-4c9d-b104-c71b65563950\") " Nov 24 07:47:34 crc kubenswrapper[4799]: I1124 07:47:34.000488 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-utilities" (OuterVolumeSpecName: "utilities") pod "3f10cb59-0728-4c9d-b104-c71b65563950" (UID: "3f10cb59-0728-4c9d-b104-c71b65563950"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:47:34 crc kubenswrapper[4799]: I1124 07:47:34.008074 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f10cb59-0728-4c9d-b104-c71b65563950-kube-api-access-wnmnx" (OuterVolumeSpecName: "kube-api-access-wnmnx") pod "3f10cb59-0728-4c9d-b104-c71b65563950" (UID: "3f10cb59-0728-4c9d-b104-c71b65563950"). InnerVolumeSpecName "kube-api-access-wnmnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:47:34 crc kubenswrapper[4799]: I1124 07:47:34.048654 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f10cb59-0728-4c9d-b104-c71b65563950" (UID: "3f10cb59-0728-4c9d-b104-c71b65563950"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:47:34 crc kubenswrapper[4799]: I1124 07:47:34.099810 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnmnx\" (UniqueName: \"kubernetes.io/projected/3f10cb59-0728-4c9d-b104-c71b65563950-kube-api-access-wnmnx\") on node \"crc\" DevicePath \"\"" Nov 24 07:47:34 crc kubenswrapper[4799]: I1124 07:47:34.099869 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:47:34 crc kubenswrapper[4799]: I1124 07:47:34.099899 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f10cb59-0728-4c9d-b104-c71b65563950-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:47:34 crc kubenswrapper[4799]: I1124 07:47:34.921783 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cg96" Nov 24 07:47:34 crc kubenswrapper[4799]: I1124 07:47:34.980328 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2cg96"] Nov 24 07:47:34 crc kubenswrapper[4799]: I1124 07:47:34.986666 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2cg96"] Nov 24 07:47:35 crc kubenswrapper[4799]: I1124 07:47:35.633308 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:47:35 crc kubenswrapper[4799]: E1124 07:47:35.633678 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:47:35 crc kubenswrapper[4799]: I1124 07:47:35.645309 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f10cb59-0728-4c9d-b104-c71b65563950" path="/var/lib/kubelet/pods/3f10cb59-0728-4c9d-b104-c71b65563950/volumes" Nov 24 07:47:46 crc kubenswrapper[4799]: I1124 07:47:46.629083 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:47:46 crc kubenswrapper[4799]: E1124 07:47:46.629998 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:48:01 crc kubenswrapper[4799]: I1124 07:48:01.628657 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:48:01 crc kubenswrapper[4799]: E1124 07:48:01.629412 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:48:16 crc kubenswrapper[4799]: I1124 07:48:16.628997 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:48:16 crc kubenswrapper[4799]: E1124 07:48:16.629873 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:48:29 crc kubenswrapper[4799]: I1124 07:48:29.628437 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:48:29 crc kubenswrapper[4799]: E1124 07:48:29.629684 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:48:43 crc kubenswrapper[4799]: I1124 07:48:43.628534 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:48:43 crc kubenswrapper[4799]: E1124 07:48:43.629648 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:48:55 crc kubenswrapper[4799]: I1124 07:48:55.636656 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:48:55 crc kubenswrapper[4799]: E1124 07:48:55.637917 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:49:06 crc kubenswrapper[4799]: I1124 07:49:06.627798 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:49:06 crc kubenswrapper[4799]: E1124 07:49:06.629009 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:49:20 crc kubenswrapper[4799]: I1124 07:49:20.628797 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:49:20 crc kubenswrapper[4799]: E1124 07:49:20.629811 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:49:34 crc kubenswrapper[4799]: I1124 07:49:34.629224 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:49:34 crc kubenswrapper[4799]: E1124 07:49:34.630410 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:49:45 crc kubenswrapper[4799]: I1124 07:49:45.637200 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:49:45 crc kubenswrapper[4799]: E1124 07:49:45.638401 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:50:00 crc kubenswrapper[4799]: I1124 07:50:00.628831 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:50:00 crc kubenswrapper[4799]: E1124 07:50:00.629836 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:50:14 crc kubenswrapper[4799]: I1124 07:50:14.628034 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:50:14 crc kubenswrapper[4799]: E1124 07:50:14.628830 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:50:29 crc kubenswrapper[4799]: I1124 07:50:29.629172 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:50:29 crc kubenswrapper[4799]: E1124 07:50:29.630088 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:50:43 crc kubenswrapper[4799]: I1124 07:50:43.629315 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:50:43 crc kubenswrapper[4799]: E1124 07:50:43.630464 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:50:58 crc kubenswrapper[4799]: I1124 07:50:58.628778 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:50:58 crc kubenswrapper[4799]: E1124 07:50:58.630019 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:51:12 crc kubenswrapper[4799]: I1124 07:51:12.629099 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:51:12 crc kubenswrapper[4799]: E1124 07:51:12.629960 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:51:26 crc kubenswrapper[4799]: I1124 07:51:26.631141 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:51:26 crc kubenswrapper[4799]: E1124 07:51:26.632449 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:51:38 crc kubenswrapper[4799]: I1124 07:51:38.628749 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:51:38 crc kubenswrapper[4799]: E1124 07:51:38.629926 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:51:50 crc kubenswrapper[4799]: I1124 07:51:50.628446 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:51:50 crc kubenswrapper[4799]: E1124 07:51:50.629691 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:52:03 crc kubenswrapper[4799]: I1124 07:52:03.629210 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:52:03 crc kubenswrapper[4799]: E1124 07:52:03.630513 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.387155 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zt9s8"] Nov 24 07:52:09 crc kubenswrapper[4799]: E1124 07:52:09.388062 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f10cb59-0728-4c9d-b104-c71b65563950" containerName="extract-content" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.388078 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f10cb59-0728-4c9d-b104-c71b65563950" containerName="extract-content" Nov 24 07:52:09 crc kubenswrapper[4799]: E1124 07:52:09.388112 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f10cb59-0728-4c9d-b104-c71b65563950" containerName="extract-utilities" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.388120 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f10cb59-0728-4c9d-b104-c71b65563950" containerName="extract-utilities" Nov 24 07:52:09 crc kubenswrapper[4799]: E1124 07:52:09.388138 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f10cb59-0728-4c9d-b104-c71b65563950" containerName="registry-server" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.388147 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f10cb59-0728-4c9d-b104-c71b65563950" containerName="registry-server" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.388327 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f10cb59-0728-4c9d-b104-c71b65563950" containerName="registry-server" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.389822 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.407464 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt9s8"] Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.537203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-catalog-content\") pod \"redhat-marketplace-zt9s8\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.537711 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcxrl\" (UniqueName: \"kubernetes.io/projected/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-kube-api-access-tcxrl\") pod \"redhat-marketplace-zt9s8\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.538032 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-utilities\") pod \"redhat-marketplace-zt9s8\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.639093 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-catalog-content\") pod \"redhat-marketplace-zt9s8\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.639159 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcxrl\" (UniqueName: \"kubernetes.io/projected/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-kube-api-access-tcxrl\") pod \"redhat-marketplace-zt9s8\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.639193 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-utilities\") pod \"redhat-marketplace-zt9s8\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.639617 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-utilities\") pod \"redhat-marketplace-zt9s8\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.639817 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-catalog-content\") pod \"redhat-marketplace-zt9s8\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.664418 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcxrl\" (UniqueName: \"kubernetes.io/projected/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-kube-api-access-tcxrl\") pod \"redhat-marketplace-zt9s8\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:09 crc kubenswrapper[4799]: I1124 07:52:09.722463 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:10 crc kubenswrapper[4799]: I1124 07:52:10.156480 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt9s8"] Nov 24 07:52:10 crc kubenswrapper[4799]: I1124 07:52:10.531328 4799 generic.go:334] "Generic (PLEG): container finished" podID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" containerID="8841922fea8ad92bfaca014c0f79ba2c1246af04c7f14dcbf19f5fd1932cf3de" exitCode=0 Nov 24 07:52:10 crc kubenswrapper[4799]: I1124 07:52:10.531372 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt9s8" event={"ID":"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f","Type":"ContainerDied","Data":"8841922fea8ad92bfaca014c0f79ba2c1246af04c7f14dcbf19f5fd1932cf3de"} Nov 24 07:52:10 crc kubenswrapper[4799]: I1124 07:52:10.531403 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt9s8" event={"ID":"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f","Type":"ContainerStarted","Data":"cc413e495e6c3fa642a2996bb98d709ee9de3e6f3657a0a5c44f3399db70ca15"} Nov 24 07:52:10 crc kubenswrapper[4799]: I1124 07:52:10.534163 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:52:11 crc kubenswrapper[4799]: I1124 07:52:11.546439 4799 generic.go:334] "Generic (PLEG): container finished" podID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" containerID="a0ff4813b657c430e8eb2409ef403b340899cf0fe8f7549c1b0c14bcfe8d3e15" exitCode=0 Nov 24 07:52:11 crc kubenswrapper[4799]: I1124 07:52:11.546531 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt9s8" event={"ID":"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f","Type":"ContainerDied","Data":"a0ff4813b657c430e8eb2409ef403b340899cf0fe8f7549c1b0c14bcfe8d3e15"} Nov 24 07:52:12 crc kubenswrapper[4799]: I1124 07:52:12.559283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt9s8" event={"ID":"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f","Type":"ContainerStarted","Data":"6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6"} Nov 24 07:52:12 crc kubenswrapper[4799]: I1124 07:52:12.587102 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zt9s8" podStartSLOduration=2.140473541 podStartE2EDuration="3.587079816s" podCreationTimestamp="2025-11-24 07:52:09 +0000 UTC" firstStartedPulling="2025-11-24 07:52:10.533920296 +0000 UTC m=+3876.189902780" lastFinishedPulling="2025-11-24 07:52:11.980526541 +0000 UTC m=+3877.636509055" observedRunningTime="2025-11-24 07:52:12.586180301 +0000 UTC m=+3878.242162795" watchObservedRunningTime="2025-11-24 07:52:12.587079816 +0000 UTC m=+3878.243062300" Nov 24 07:52:16 crc kubenswrapper[4799]: I1124 07:52:16.629138 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:52:16 crc kubenswrapper[4799]: E1124 07:52:16.630297 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:52:19 crc kubenswrapper[4799]: I1124 07:52:19.728497 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:19 crc kubenswrapper[4799]: I1124 07:52:19.729170 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:19 crc kubenswrapper[4799]: I1124 07:52:19.937835 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:20 crc kubenswrapper[4799]: I1124 07:52:20.712055 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:20 crc kubenswrapper[4799]: I1124 07:52:20.786316 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt9s8"] Nov 24 07:52:22 crc kubenswrapper[4799]: I1124 07:52:22.650415 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zt9s8" podUID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" containerName="registry-server" containerID="cri-o://6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6" gracePeriod=2 Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.157872 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.266112 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcxrl\" (UniqueName: \"kubernetes.io/projected/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-kube-api-access-tcxrl\") pod \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.266279 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-catalog-content\") pod \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.266315 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-utilities\") pod \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\" (UID: \"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f\") " Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.267666 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-utilities" (OuterVolumeSpecName: "utilities") pod "26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" (UID: "26ee63e4-cdd0-4b7c-ba92-296a7a0f140f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.275120 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-kube-api-access-tcxrl" (OuterVolumeSpecName: "kube-api-access-tcxrl") pod "26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" (UID: "26ee63e4-cdd0-4b7c-ba92-296a7a0f140f"). InnerVolumeSpecName "kube-api-access-tcxrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.293188 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" (UID: "26ee63e4-cdd0-4b7c-ba92-296a7a0f140f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.368433 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcxrl\" (UniqueName: \"kubernetes.io/projected/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-kube-api-access-tcxrl\") on node \"crc\" DevicePath \"\"" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.368470 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.368479 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.663287 4799 generic.go:334] "Generic (PLEG): container finished" podID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" containerID="6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6" exitCode=0 Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.663369 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zt9s8" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.663409 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt9s8" event={"ID":"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f","Type":"ContainerDied","Data":"6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6"} Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.664314 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt9s8" event={"ID":"26ee63e4-cdd0-4b7c-ba92-296a7a0f140f","Type":"ContainerDied","Data":"cc413e495e6c3fa642a2996bb98d709ee9de3e6f3657a0a5c44f3399db70ca15"} Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.664353 4799 scope.go:117] "RemoveContainer" containerID="6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.712582 4799 scope.go:117] "RemoveContainer" containerID="a0ff4813b657c430e8eb2409ef403b340899cf0fe8f7549c1b0c14bcfe8d3e15" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.715351 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt9s8"] Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.727362 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt9s8"] Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.747450 4799 scope.go:117] "RemoveContainer" containerID="8841922fea8ad92bfaca014c0f79ba2c1246af04c7f14dcbf19f5fd1932cf3de" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.778397 4799 scope.go:117] "RemoveContainer" containerID="6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6" Nov 24 07:52:23 crc kubenswrapper[4799]: E1124 07:52:23.779102 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6\": container with ID starting with 6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6 not found: ID does not exist" containerID="6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.779148 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6"} err="failed to get container status \"6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6\": rpc error: code = NotFound desc = could not find container \"6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6\": container with ID starting with 6826a6294a6bf132eec2ebd3ef65cd18812e2d91f735e035c7ae5e5070b5b0a6 not found: ID does not exist" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.779179 4799 scope.go:117] "RemoveContainer" containerID="a0ff4813b657c430e8eb2409ef403b340899cf0fe8f7549c1b0c14bcfe8d3e15" Nov 24 07:52:23 crc kubenswrapper[4799]: E1124 07:52:23.779563 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0ff4813b657c430e8eb2409ef403b340899cf0fe8f7549c1b0c14bcfe8d3e15\": container with ID starting with a0ff4813b657c430e8eb2409ef403b340899cf0fe8f7549c1b0c14bcfe8d3e15 not found: ID does not exist" containerID="a0ff4813b657c430e8eb2409ef403b340899cf0fe8f7549c1b0c14bcfe8d3e15" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.779592 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ff4813b657c430e8eb2409ef403b340899cf0fe8f7549c1b0c14bcfe8d3e15"} err="failed to get container status \"a0ff4813b657c430e8eb2409ef403b340899cf0fe8f7549c1b0c14bcfe8d3e15\": rpc error: code = NotFound desc = could not find container \"a0ff4813b657c430e8eb2409ef403b340899cf0fe8f7549c1b0c14bcfe8d3e15\": container with ID starting with a0ff4813b657c430e8eb2409ef403b340899cf0fe8f7549c1b0c14bcfe8d3e15 not found: ID does not exist" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.779614 4799 scope.go:117] "RemoveContainer" containerID="8841922fea8ad92bfaca014c0f79ba2c1246af04c7f14dcbf19f5fd1932cf3de" Nov 24 07:52:23 crc kubenswrapper[4799]: E1124 07:52:23.780060 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8841922fea8ad92bfaca014c0f79ba2c1246af04c7f14dcbf19f5fd1932cf3de\": container with ID starting with 8841922fea8ad92bfaca014c0f79ba2c1246af04c7f14dcbf19f5fd1932cf3de not found: ID does not exist" containerID="8841922fea8ad92bfaca014c0f79ba2c1246af04c7f14dcbf19f5fd1932cf3de" Nov 24 07:52:23 crc kubenswrapper[4799]: I1124 07:52:23.780118 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8841922fea8ad92bfaca014c0f79ba2c1246af04c7f14dcbf19f5fd1932cf3de"} err="failed to get container status \"8841922fea8ad92bfaca014c0f79ba2c1246af04c7f14dcbf19f5fd1932cf3de\": rpc error: code = NotFound desc = could not find container \"8841922fea8ad92bfaca014c0f79ba2c1246af04c7f14dcbf19f5fd1932cf3de\": container with ID starting with 8841922fea8ad92bfaca014c0f79ba2c1246af04c7f14dcbf19f5fd1932cf3de not found: ID does not exist" Nov 24 07:52:25 crc kubenswrapper[4799]: I1124 07:52:25.644601 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" path="/var/lib/kubelet/pods/26ee63e4-cdd0-4b7c-ba92-296a7a0f140f/volumes" Nov 24 07:52:31 crc kubenswrapper[4799]: I1124 07:52:31.629971 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:52:32 crc kubenswrapper[4799]: I1124 07:52:32.761776 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"79a567e5468d72ae31fb5bce03a697890201831b25f9fd884a4780275b786543"} Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.304510 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-pvnqk"] Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.313753 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-pvnqk"] Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.479257 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-wpvf8"] Nov 24 07:53:10 crc kubenswrapper[4799]: E1124 07:53:10.479660 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" containerName="registry-server" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.479678 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" containerName="registry-server" Nov 24 07:53:10 crc kubenswrapper[4799]: E1124 07:53:10.479702 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" containerName="extract-utilities" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.479712 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" containerName="extract-utilities" Nov 24 07:53:10 crc kubenswrapper[4799]: E1124 07:53:10.479739 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" containerName="extract-content" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.479748 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" containerName="extract-content" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.479955 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="26ee63e4-cdd0-4b7c-ba92-296a7a0f140f" containerName="registry-server" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.480508 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.492285 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.492391 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.492397 4799 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-2smct" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.492483 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.499600 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-wpvf8"] Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.538510 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/3f3f0de1-1026-4dab-b5b4-5792b212bd45-node-mnt\") pod \"crc-storage-crc-wpvf8\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.538601 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/3f3f0de1-1026-4dab-b5b4-5792b212bd45-crc-storage\") pod \"crc-storage-crc-wpvf8\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.538681 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m29ld\" (UniqueName: \"kubernetes.io/projected/3f3f0de1-1026-4dab-b5b4-5792b212bd45-kube-api-access-m29ld\") pod \"crc-storage-crc-wpvf8\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.640383 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m29ld\" (UniqueName: \"kubernetes.io/projected/3f3f0de1-1026-4dab-b5b4-5792b212bd45-kube-api-access-m29ld\") pod \"crc-storage-crc-wpvf8\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.640513 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/3f3f0de1-1026-4dab-b5b4-5792b212bd45-node-mnt\") pod \"crc-storage-crc-wpvf8\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.640691 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/3f3f0de1-1026-4dab-b5b4-5792b212bd45-crc-storage\") pod \"crc-storage-crc-wpvf8\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.641130 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/3f3f0de1-1026-4dab-b5b4-5792b212bd45-node-mnt\") pod \"crc-storage-crc-wpvf8\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.642253 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/3f3f0de1-1026-4dab-b5b4-5792b212bd45-crc-storage\") pod \"crc-storage-crc-wpvf8\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.672072 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m29ld\" (UniqueName: \"kubernetes.io/projected/3f3f0de1-1026-4dab-b5b4-5792b212bd45-kube-api-access-m29ld\") pod \"crc-storage-crc-wpvf8\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:10 crc kubenswrapper[4799]: I1124 07:53:10.818376 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:11 crc kubenswrapper[4799]: I1124 07:53:11.078572 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-wpvf8"] Nov 24 07:53:11 crc kubenswrapper[4799]: I1124 07:53:11.180509 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wpvf8" event={"ID":"3f3f0de1-1026-4dab-b5b4-5792b212bd45","Type":"ContainerStarted","Data":"eca4fac88e949c5de379230dfeba05de887f2008759d9fc59235a0393016f812"} Nov 24 07:53:11 crc kubenswrapper[4799]: I1124 07:53:11.643410 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65589a34-a1fb-49cb-ba7b-57da57096604" path="/var/lib/kubelet/pods/65589a34-a1fb-49cb-ba7b-57da57096604/volumes" Nov 24 07:53:12 crc kubenswrapper[4799]: I1124 07:53:12.192633 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wpvf8" event={"ID":"3f3f0de1-1026-4dab-b5b4-5792b212bd45","Type":"ContainerStarted","Data":"2b06ede5a007b86e84f723b97557a301b6f40e96cb94c638f67ade122b9bfd2a"} Nov 24 07:53:12 crc kubenswrapper[4799]: I1124 07:53:12.225481 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-wpvf8" podStartSLOduration=1.706045129 podStartE2EDuration="2.225450842s" podCreationTimestamp="2025-11-24 07:53:10 +0000 UTC" firstStartedPulling="2025-11-24 07:53:11.0863571 +0000 UTC m=+3936.742339574" lastFinishedPulling="2025-11-24 07:53:11.605762773 +0000 UTC m=+3937.261745287" observedRunningTime="2025-11-24 07:53:12.212664239 +0000 UTC m=+3937.868646763" watchObservedRunningTime="2025-11-24 07:53:12.225450842 +0000 UTC m=+3937.881433356" Nov 24 07:53:13 crc kubenswrapper[4799]: I1124 07:53:13.206372 4799 generic.go:334] "Generic (PLEG): container finished" podID="3f3f0de1-1026-4dab-b5b4-5792b212bd45" containerID="2b06ede5a007b86e84f723b97557a301b6f40e96cb94c638f67ade122b9bfd2a" exitCode=0 Nov 24 07:53:13 crc kubenswrapper[4799]: I1124 07:53:13.207007 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wpvf8" event={"ID":"3f3f0de1-1026-4dab-b5b4-5792b212bd45","Type":"ContainerDied","Data":"2b06ede5a007b86e84f723b97557a301b6f40e96cb94c638f67ade122b9bfd2a"} Nov 24 07:53:14 crc kubenswrapper[4799]: I1124 07:53:14.639151 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:14 crc kubenswrapper[4799]: I1124 07:53:14.806316 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/3f3f0de1-1026-4dab-b5b4-5792b212bd45-crc-storage\") pod \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " Nov 24 07:53:14 crc kubenswrapper[4799]: I1124 07:53:14.806406 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m29ld\" (UniqueName: \"kubernetes.io/projected/3f3f0de1-1026-4dab-b5b4-5792b212bd45-kube-api-access-m29ld\") pod \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " Nov 24 07:53:14 crc kubenswrapper[4799]: I1124 07:53:14.806557 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/3f3f0de1-1026-4dab-b5b4-5792b212bd45-node-mnt\") pod \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\" (UID: \"3f3f0de1-1026-4dab-b5b4-5792b212bd45\") " Nov 24 07:53:14 crc kubenswrapper[4799]: I1124 07:53:14.807191 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f3f0de1-1026-4dab-b5b4-5792b212bd45-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "3f3f0de1-1026-4dab-b5b4-5792b212bd45" (UID: "3f3f0de1-1026-4dab-b5b4-5792b212bd45"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:53:14 crc kubenswrapper[4799]: I1124 07:53:14.827390 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f3f0de1-1026-4dab-b5b4-5792b212bd45-kube-api-access-m29ld" (OuterVolumeSpecName: "kube-api-access-m29ld") pod "3f3f0de1-1026-4dab-b5b4-5792b212bd45" (UID: "3f3f0de1-1026-4dab-b5b4-5792b212bd45"). InnerVolumeSpecName "kube-api-access-m29ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:53:14 crc kubenswrapper[4799]: I1124 07:53:14.841311 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f3f0de1-1026-4dab-b5b4-5792b212bd45-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "3f3f0de1-1026-4dab-b5b4-5792b212bd45" (UID: "3f3f0de1-1026-4dab-b5b4-5792b212bd45"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:53:14 crc kubenswrapper[4799]: I1124 07:53:14.908592 4799 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/3f3f0de1-1026-4dab-b5b4-5792b212bd45-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:14 crc kubenswrapper[4799]: I1124 07:53:14.908919 4799 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/3f3f0de1-1026-4dab-b5b4-5792b212bd45-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:14 crc kubenswrapper[4799]: I1124 07:53:14.909035 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m29ld\" (UniqueName: \"kubernetes.io/projected/3f3f0de1-1026-4dab-b5b4-5792b212bd45-kube-api-access-m29ld\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:15 crc kubenswrapper[4799]: I1124 07:53:15.229181 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wpvf8" event={"ID":"3f3f0de1-1026-4dab-b5b4-5792b212bd45","Type":"ContainerDied","Data":"eca4fac88e949c5de379230dfeba05de887f2008759d9fc59235a0393016f812"} Nov 24 07:53:15 crc kubenswrapper[4799]: I1124 07:53:15.229258 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eca4fac88e949c5de379230dfeba05de887f2008759d9fc59235a0393016f812" Nov 24 07:53:15 crc kubenswrapper[4799]: I1124 07:53:15.229276 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wpvf8" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.586886 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-wpvf8"] Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.595000 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-wpvf8"] Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.770931 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-bnlc9"] Nov 24 07:53:16 crc kubenswrapper[4799]: E1124 07:53:16.771415 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f3f0de1-1026-4dab-b5b4-5792b212bd45" containerName="storage" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.771446 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f3f0de1-1026-4dab-b5b4-5792b212bd45" containerName="storage" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.771664 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f3f0de1-1026-4dab-b5b4-5792b212bd45" containerName="storage" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.772470 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.774985 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.775689 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.776135 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.776489 4799 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-2smct" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.783677 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-bnlc9"] Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.857993 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/2bf3f312-d410-4400-bf45-9e4b170d3cda-crc-storage\") pod \"crc-storage-crc-bnlc9\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.858376 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/2bf3f312-d410-4400-bf45-9e4b170d3cda-node-mnt\") pod \"crc-storage-crc-bnlc9\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.858512 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nks9v\" (UniqueName: \"kubernetes.io/projected/2bf3f312-d410-4400-bf45-9e4b170d3cda-kube-api-access-nks9v\") pod \"crc-storage-crc-bnlc9\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.959940 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/2bf3f312-d410-4400-bf45-9e4b170d3cda-crc-storage\") pod \"crc-storage-crc-bnlc9\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.960407 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/2bf3f312-d410-4400-bf45-9e4b170d3cda-node-mnt\") pod \"crc-storage-crc-bnlc9\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.960521 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nks9v\" (UniqueName: \"kubernetes.io/projected/2bf3f312-d410-4400-bf45-9e4b170d3cda-kube-api-access-nks9v\") pod \"crc-storage-crc-bnlc9\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.960708 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/2bf3f312-d410-4400-bf45-9e4b170d3cda-node-mnt\") pod \"crc-storage-crc-bnlc9\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.960829 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/2bf3f312-d410-4400-bf45-9e4b170d3cda-crc-storage\") pod \"crc-storage-crc-bnlc9\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:16 crc kubenswrapper[4799]: I1124 07:53:16.984727 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nks9v\" (UniqueName: \"kubernetes.io/projected/2bf3f312-d410-4400-bf45-9e4b170d3cda-kube-api-access-nks9v\") pod \"crc-storage-crc-bnlc9\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:17 crc kubenswrapper[4799]: I1124 07:53:17.103351 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:17 crc kubenswrapper[4799]: I1124 07:53:17.528391 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-bnlc9"] Nov 24 07:53:17 crc kubenswrapper[4799]: I1124 07:53:17.639231 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f3f0de1-1026-4dab-b5b4-5792b212bd45" path="/var/lib/kubelet/pods/3f3f0de1-1026-4dab-b5b4-5792b212bd45/volumes" Nov 24 07:53:18 crc kubenswrapper[4799]: I1124 07:53:18.256386 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-bnlc9" event={"ID":"2bf3f312-d410-4400-bf45-9e4b170d3cda","Type":"ContainerStarted","Data":"21c58fd1c1eccfc8de03b52f9e863de5a72ad39c067461e0285d3fdf6b215d2c"} Nov 24 07:53:18 crc kubenswrapper[4799]: E1124 07:53:18.495056 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bf3f312_d410_4400_bf45_9e4b170d3cda.slice/crio-e7229c8564ef78de67823dc85d2303265cfbdfab0c6e5a21b5c7e7a5e7f51894.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bf3f312_d410_4400_bf45_9e4b170d3cda.slice/crio-conmon-e7229c8564ef78de67823dc85d2303265cfbdfab0c6e5a21b5c7e7a5e7f51894.scope\": RecentStats: unable to find data in memory cache]" Nov 24 07:53:19 crc kubenswrapper[4799]: I1124 07:53:19.273325 4799 generic.go:334] "Generic (PLEG): container finished" podID="2bf3f312-d410-4400-bf45-9e4b170d3cda" containerID="e7229c8564ef78de67823dc85d2303265cfbdfab0c6e5a21b5c7e7a5e7f51894" exitCode=0 Nov 24 07:53:19 crc kubenswrapper[4799]: I1124 07:53:19.273436 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-bnlc9" event={"ID":"2bf3f312-d410-4400-bf45-9e4b170d3cda","Type":"ContainerDied","Data":"e7229c8564ef78de67823dc85d2303265cfbdfab0c6e5a21b5c7e7a5e7f51894"} Nov 24 07:53:20 crc kubenswrapper[4799]: I1124 07:53:20.670771 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:20 crc kubenswrapper[4799]: I1124 07:53:20.720188 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nks9v\" (UniqueName: \"kubernetes.io/projected/2bf3f312-d410-4400-bf45-9e4b170d3cda-kube-api-access-nks9v\") pod \"2bf3f312-d410-4400-bf45-9e4b170d3cda\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " Nov 24 07:53:20 crc kubenswrapper[4799]: I1124 07:53:20.720317 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/2bf3f312-d410-4400-bf45-9e4b170d3cda-node-mnt\") pod \"2bf3f312-d410-4400-bf45-9e4b170d3cda\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " Nov 24 07:53:20 crc kubenswrapper[4799]: I1124 07:53:20.720358 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/2bf3f312-d410-4400-bf45-9e4b170d3cda-crc-storage\") pod \"2bf3f312-d410-4400-bf45-9e4b170d3cda\" (UID: \"2bf3f312-d410-4400-bf45-9e4b170d3cda\") " Nov 24 07:53:20 crc kubenswrapper[4799]: I1124 07:53:20.720448 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2bf3f312-d410-4400-bf45-9e4b170d3cda-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "2bf3f312-d410-4400-bf45-9e4b170d3cda" (UID: "2bf3f312-d410-4400-bf45-9e4b170d3cda"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 07:53:20 crc kubenswrapper[4799]: I1124 07:53:20.722029 4799 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/2bf3f312-d410-4400-bf45-9e4b170d3cda-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:20 crc kubenswrapper[4799]: I1124 07:53:20.756427 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bf3f312-d410-4400-bf45-9e4b170d3cda-kube-api-access-nks9v" (OuterVolumeSpecName: "kube-api-access-nks9v") pod "2bf3f312-d410-4400-bf45-9e4b170d3cda" (UID: "2bf3f312-d410-4400-bf45-9e4b170d3cda"). InnerVolumeSpecName "kube-api-access-nks9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:53:20 crc kubenswrapper[4799]: I1124 07:53:20.781697 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bf3f312-d410-4400-bf45-9e4b170d3cda-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "2bf3f312-d410-4400-bf45-9e4b170d3cda" (UID: "2bf3f312-d410-4400-bf45-9e4b170d3cda"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:53:20 crc kubenswrapper[4799]: I1124 07:53:20.823395 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nks9v\" (UniqueName: \"kubernetes.io/projected/2bf3f312-d410-4400-bf45-9e4b170d3cda-kube-api-access-nks9v\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:20 crc kubenswrapper[4799]: I1124 07:53:20.823430 4799 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/2bf3f312-d410-4400-bf45-9e4b170d3cda-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 24 07:53:21 crc kubenswrapper[4799]: I1124 07:53:21.297650 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-bnlc9" event={"ID":"2bf3f312-d410-4400-bf45-9e4b170d3cda","Type":"ContainerDied","Data":"21c58fd1c1eccfc8de03b52f9e863de5a72ad39c067461e0285d3fdf6b215d2c"} Nov 24 07:53:21 crc kubenswrapper[4799]: I1124 07:53:21.297740 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21c58fd1c1eccfc8de03b52f9e863de5a72ad39c067461e0285d3fdf6b215d2c" Nov 24 07:53:21 crc kubenswrapper[4799]: I1124 07:53:21.297814 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-bnlc9" Nov 24 07:53:53 crc kubenswrapper[4799]: I1124 07:53:53.782994 4799 scope.go:117] "RemoveContainer" containerID="f3b7bfe56f907040c8123c075eae2d219a2ae33843b9e40f05d41179e21aa19c" Nov 24 07:53:53 crc kubenswrapper[4799]: I1124 07:53:53.838052 4799 scope.go:117] "RemoveContainer" containerID="3cec9c21bcd7ccf8f36fc17ed0a663c1fa4b6692d475d6510b1a131a500d1c84" Nov 24 07:53:53 crc kubenswrapper[4799]: I1124 07:53:53.875255 4799 scope.go:117] "RemoveContainer" containerID="eedc52f33720d3cf8d8151e4505ec6de3450715f191355ee06eaa8418217f794" Nov 24 07:53:53 crc kubenswrapper[4799]: I1124 07:53:53.905051 4799 scope.go:117] "RemoveContainer" containerID="86139a0afdb7e9b1522fc94ff46ebbc967827ec9eb2832f5df2718834704705d" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.195926 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vczg9"] Nov 24 07:53:57 crc kubenswrapper[4799]: E1124 07:53:57.196716 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf3f312-d410-4400-bf45-9e4b170d3cda" containerName="storage" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.196738 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf3f312-d410-4400-bf45-9e4b170d3cda" containerName="storage" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.197021 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bf3f312-d410-4400-bf45-9e4b170d3cda" containerName="storage" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.198749 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.220373 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vczg9"] Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.351802 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-utilities\") pod \"community-operators-vczg9\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.351951 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2ldr\" (UniqueName: \"kubernetes.io/projected/f4ef9113-c417-420c-b1e6-00c6af35351e-kube-api-access-d2ldr\") pod \"community-operators-vczg9\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.352003 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-catalog-content\") pod \"community-operators-vczg9\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.453243 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-utilities\") pod \"community-operators-vczg9\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.453341 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2ldr\" (UniqueName: \"kubernetes.io/projected/f4ef9113-c417-420c-b1e6-00c6af35351e-kube-api-access-d2ldr\") pod \"community-operators-vczg9\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.453395 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-catalog-content\") pod \"community-operators-vczg9\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.454028 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-utilities\") pod \"community-operators-vczg9\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.454043 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-catalog-content\") pod \"community-operators-vczg9\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.477201 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2ldr\" (UniqueName: \"kubernetes.io/projected/f4ef9113-c417-420c-b1e6-00c6af35351e-kube-api-access-d2ldr\") pod \"community-operators-vczg9\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.546382 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:53:57 crc kubenswrapper[4799]: I1124 07:53:57.801388 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vczg9"] Nov 24 07:53:58 crc kubenswrapper[4799]: I1124 07:53:58.673937 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4ef9113-c417-420c-b1e6-00c6af35351e" containerID="6aeca99ae8b6cf080668a6c0519a2d402ecb13f0f438544ff3cd0bb2018c75d8" exitCode=0 Nov 24 07:53:58 crc kubenswrapper[4799]: I1124 07:53:58.674030 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vczg9" event={"ID":"f4ef9113-c417-420c-b1e6-00c6af35351e","Type":"ContainerDied","Data":"6aeca99ae8b6cf080668a6c0519a2d402ecb13f0f438544ff3cd0bb2018c75d8"} Nov 24 07:53:58 crc kubenswrapper[4799]: I1124 07:53:58.674209 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vczg9" event={"ID":"f4ef9113-c417-420c-b1e6-00c6af35351e","Type":"ContainerStarted","Data":"22a7a39c2498eb88290a2c27c51bca1a68b5863c0c3b510beb9e6260e32e4749"} Nov 24 07:53:59 crc kubenswrapper[4799]: I1124 07:53:59.688066 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vczg9" event={"ID":"f4ef9113-c417-420c-b1e6-00c6af35351e","Type":"ContainerStarted","Data":"510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25"} Nov 24 07:54:00 crc kubenswrapper[4799]: I1124 07:54:00.701920 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4ef9113-c417-420c-b1e6-00c6af35351e" containerID="510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25" exitCode=0 Nov 24 07:54:00 crc kubenswrapper[4799]: I1124 07:54:00.701990 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vczg9" event={"ID":"f4ef9113-c417-420c-b1e6-00c6af35351e","Type":"ContainerDied","Data":"510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25"} Nov 24 07:54:01 crc kubenswrapper[4799]: I1124 07:54:01.716199 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vczg9" event={"ID":"f4ef9113-c417-420c-b1e6-00c6af35351e","Type":"ContainerStarted","Data":"b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134"} Nov 24 07:54:01 crc kubenswrapper[4799]: I1124 07:54:01.743041 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vczg9" podStartSLOduration=2.282660763 podStartE2EDuration="4.743015233s" podCreationTimestamp="2025-11-24 07:53:57 +0000 UTC" firstStartedPulling="2025-11-24 07:53:58.676134838 +0000 UTC m=+3984.332117312" lastFinishedPulling="2025-11-24 07:54:01.136489278 +0000 UTC m=+3986.792471782" observedRunningTime="2025-11-24 07:54:01.741334715 +0000 UTC m=+3987.397317209" watchObservedRunningTime="2025-11-24 07:54:01.743015233 +0000 UTC m=+3987.398997717" Nov 24 07:54:07 crc kubenswrapper[4799]: I1124 07:54:07.547127 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:54:07 crc kubenswrapper[4799]: I1124 07:54:07.548192 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:54:07 crc kubenswrapper[4799]: I1124 07:54:07.613228 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:54:07 crc kubenswrapper[4799]: I1124 07:54:07.806223 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:54:07 crc kubenswrapper[4799]: I1124 07:54:07.850416 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vczg9"] Nov 24 07:54:09 crc kubenswrapper[4799]: I1124 07:54:09.794406 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vczg9" podUID="f4ef9113-c417-420c-b1e6-00c6af35351e" containerName="registry-server" containerID="cri-o://b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134" gracePeriod=2 Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.308592 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.360953 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2ldr\" (UniqueName: \"kubernetes.io/projected/f4ef9113-c417-420c-b1e6-00c6af35351e-kube-api-access-d2ldr\") pod \"f4ef9113-c417-420c-b1e6-00c6af35351e\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.361036 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-catalog-content\") pod \"f4ef9113-c417-420c-b1e6-00c6af35351e\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.361246 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-utilities\") pod \"f4ef9113-c417-420c-b1e6-00c6af35351e\" (UID: \"f4ef9113-c417-420c-b1e6-00c6af35351e\") " Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.363471 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-utilities" (OuterVolumeSpecName: "utilities") pod "f4ef9113-c417-420c-b1e6-00c6af35351e" (UID: "f4ef9113-c417-420c-b1e6-00c6af35351e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.369042 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4ef9113-c417-420c-b1e6-00c6af35351e-kube-api-access-d2ldr" (OuterVolumeSpecName: "kube-api-access-d2ldr") pod "f4ef9113-c417-420c-b1e6-00c6af35351e" (UID: "f4ef9113-c417-420c-b1e6-00c6af35351e"). InnerVolumeSpecName "kube-api-access-d2ldr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.421192 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4ef9113-c417-420c-b1e6-00c6af35351e" (UID: "f4ef9113-c417-420c-b1e6-00c6af35351e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.463548 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2ldr\" (UniqueName: \"kubernetes.io/projected/f4ef9113-c417-420c-b1e6-00c6af35351e-kube-api-access-d2ldr\") on node \"crc\" DevicePath \"\"" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.463611 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.463629 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4ef9113-c417-420c-b1e6-00c6af35351e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.808744 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4ef9113-c417-420c-b1e6-00c6af35351e" containerID="b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134" exitCode=0 Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.808819 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vczg9" event={"ID":"f4ef9113-c417-420c-b1e6-00c6af35351e","Type":"ContainerDied","Data":"b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134"} Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.808908 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vczg9" event={"ID":"f4ef9113-c417-420c-b1e6-00c6af35351e","Type":"ContainerDied","Data":"22a7a39c2498eb88290a2c27c51bca1a68b5863c0c3b510beb9e6260e32e4749"} Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.808912 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vczg9" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.808950 4799 scope.go:117] "RemoveContainer" containerID="b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.842028 4799 scope.go:117] "RemoveContainer" containerID="510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.867420 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vczg9"] Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.874829 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vczg9"] Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.881195 4799 scope.go:117] "RemoveContainer" containerID="6aeca99ae8b6cf080668a6c0519a2d402ecb13f0f438544ff3cd0bb2018c75d8" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.904027 4799 scope.go:117] "RemoveContainer" containerID="b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134" Nov 24 07:54:10 crc kubenswrapper[4799]: E1124 07:54:10.904441 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134\": container with ID starting with b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134 not found: ID does not exist" containerID="b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.904474 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134"} err="failed to get container status \"b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134\": rpc error: code = NotFound desc = could not find container \"b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134\": container with ID starting with b9e3adb9cc3c5be19a4152728e8d9fdd0c8a11495cd9946911a93b8ad40f4134 not found: ID does not exist" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.904495 4799 scope.go:117] "RemoveContainer" containerID="510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25" Nov 24 07:54:10 crc kubenswrapper[4799]: E1124 07:54:10.904764 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25\": container with ID starting with 510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25 not found: ID does not exist" containerID="510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.904812 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25"} err="failed to get container status \"510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25\": rpc error: code = NotFound desc = could not find container \"510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25\": container with ID starting with 510ab94c4b0c40ff6dac515d180584703f3e1692521b66830ada45a973183e25 not found: ID does not exist" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.904827 4799 scope.go:117] "RemoveContainer" containerID="6aeca99ae8b6cf080668a6c0519a2d402ecb13f0f438544ff3cd0bb2018c75d8" Nov 24 07:54:10 crc kubenswrapper[4799]: E1124 07:54:10.905248 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aeca99ae8b6cf080668a6c0519a2d402ecb13f0f438544ff3cd0bb2018c75d8\": container with ID starting with 6aeca99ae8b6cf080668a6c0519a2d402ecb13f0f438544ff3cd0bb2018c75d8 not found: ID does not exist" containerID="6aeca99ae8b6cf080668a6c0519a2d402ecb13f0f438544ff3cd0bb2018c75d8" Nov 24 07:54:10 crc kubenswrapper[4799]: I1124 07:54:10.905315 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aeca99ae8b6cf080668a6c0519a2d402ecb13f0f438544ff3cd0bb2018c75d8"} err="failed to get container status \"6aeca99ae8b6cf080668a6c0519a2d402ecb13f0f438544ff3cd0bb2018c75d8\": rpc error: code = NotFound desc = could not find container \"6aeca99ae8b6cf080668a6c0519a2d402ecb13f0f438544ff3cd0bb2018c75d8\": container with ID starting with 6aeca99ae8b6cf080668a6c0519a2d402ecb13f0f438544ff3cd0bb2018c75d8 not found: ID does not exist" Nov 24 07:54:11 crc kubenswrapper[4799]: I1124 07:54:11.638621 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4ef9113-c417-420c-b1e6-00c6af35351e" path="/var/lib/kubelet/pods/f4ef9113-c417-420c-b1e6-00c6af35351e/volumes" Nov 24 07:54:50 crc kubenswrapper[4799]: I1124 07:54:50.400379 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:54:50 crc kubenswrapper[4799]: I1124 07:54:50.401116 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:55:20 crc kubenswrapper[4799]: I1124 07:55:20.400607 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:55:20 crc kubenswrapper[4799]: I1124 07:55:20.401327 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:55:50 crc kubenswrapper[4799]: I1124 07:55:50.401183 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:55:50 crc kubenswrapper[4799]: I1124 07:55:50.401838 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:55:50 crc kubenswrapper[4799]: I1124 07:55:50.401963 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:55:50 crc kubenswrapper[4799]: I1124 07:55:50.403071 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"79a567e5468d72ae31fb5bce03a697890201831b25f9fd884a4780275b786543"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:55:50 crc kubenswrapper[4799]: I1124 07:55:50.403186 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://79a567e5468d72ae31fb5bce03a697890201831b25f9fd884a4780275b786543" gracePeriod=600 Nov 24 07:55:50 crc kubenswrapper[4799]: I1124 07:55:50.854469 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="79a567e5468d72ae31fb5bce03a697890201831b25f9fd884a4780275b786543" exitCode=0 Nov 24 07:55:50 crc kubenswrapper[4799]: I1124 07:55:50.854552 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"79a567e5468d72ae31fb5bce03a697890201831b25f9fd884a4780275b786543"} Nov 24 07:55:50 crc kubenswrapper[4799]: I1124 07:55:50.854833 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323"} Nov 24 07:55:50 crc kubenswrapper[4799]: I1124 07:55:50.854898 4799 scope.go:117] "RemoveContainer" containerID="3dc12ce57ec6d563ad73a00f44a6b58dd2f7bbea80a692def84f21e205699832" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.304527 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4tfxf"] Nov 24 07:56:08 crc kubenswrapper[4799]: E1124 07:56:08.305398 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4ef9113-c417-420c-b1e6-00c6af35351e" containerName="extract-content" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.305413 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4ef9113-c417-420c-b1e6-00c6af35351e" containerName="extract-content" Nov 24 07:56:08 crc kubenswrapper[4799]: E1124 07:56:08.305426 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4ef9113-c417-420c-b1e6-00c6af35351e" containerName="extract-utilities" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.305432 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4ef9113-c417-420c-b1e6-00c6af35351e" containerName="extract-utilities" Nov 24 07:56:08 crc kubenswrapper[4799]: E1124 07:56:08.305444 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4ef9113-c417-420c-b1e6-00c6af35351e" containerName="registry-server" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.305451 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4ef9113-c417-420c-b1e6-00c6af35351e" containerName="registry-server" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.305580 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4ef9113-c417-420c-b1e6-00c6af35351e" containerName="registry-server" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.306504 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.329052 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4tfxf"] Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.405590 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-catalog-content\") pod \"redhat-operators-4tfxf\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.405728 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2v8p\" (UniqueName: \"kubernetes.io/projected/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-kube-api-access-d2v8p\") pod \"redhat-operators-4tfxf\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.405759 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-utilities\") pod \"redhat-operators-4tfxf\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.506462 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-catalog-content\") pod \"redhat-operators-4tfxf\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.506548 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2v8p\" (UniqueName: \"kubernetes.io/projected/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-kube-api-access-d2v8p\") pod \"redhat-operators-4tfxf\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.506568 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-utilities\") pod \"redhat-operators-4tfxf\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.507099 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-utilities\") pod \"redhat-operators-4tfxf\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.507361 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-catalog-content\") pod \"redhat-operators-4tfxf\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.672452 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2v8p\" (UniqueName: \"kubernetes.io/projected/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-kube-api-access-d2v8p\") pod \"redhat-operators-4tfxf\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:08 crc kubenswrapper[4799]: I1124 07:56:08.944540 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:09 crc kubenswrapper[4799]: I1124 07:56:09.434804 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4tfxf"] Nov 24 07:56:10 crc kubenswrapper[4799]: I1124 07:56:10.074111 4799 generic.go:334] "Generic (PLEG): container finished" podID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerID="0ca7495bbf2c43a9e6db5ad564e87edea3e0851ff413435d7a94b178213cd397" exitCode=0 Nov 24 07:56:10 crc kubenswrapper[4799]: I1124 07:56:10.074176 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4tfxf" event={"ID":"14a602c5-0fe1-4b5c-b15c-6fc532fd863b","Type":"ContainerDied","Data":"0ca7495bbf2c43a9e6db5ad564e87edea3e0851ff413435d7a94b178213cd397"} Nov 24 07:56:10 crc kubenswrapper[4799]: I1124 07:56:10.074665 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4tfxf" event={"ID":"14a602c5-0fe1-4b5c-b15c-6fc532fd863b","Type":"ContainerStarted","Data":"bb6bbfb15105837e265b55b66fb93ee8eb86999a05a47f515734f4c7aef25790"} Nov 24 07:56:11 crc kubenswrapper[4799]: I1124 07:56:11.087837 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4tfxf" event={"ID":"14a602c5-0fe1-4b5c-b15c-6fc532fd863b","Type":"ContainerStarted","Data":"8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854"} Nov 24 07:56:12 crc kubenswrapper[4799]: I1124 07:56:12.103344 4799 generic.go:334] "Generic (PLEG): container finished" podID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerID="8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854" exitCode=0 Nov 24 07:56:12 crc kubenswrapper[4799]: I1124 07:56:12.103409 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4tfxf" event={"ID":"14a602c5-0fe1-4b5c-b15c-6fc532fd863b","Type":"ContainerDied","Data":"8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854"} Nov 24 07:56:13 crc kubenswrapper[4799]: I1124 07:56:13.124100 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4tfxf" event={"ID":"14a602c5-0fe1-4b5c-b15c-6fc532fd863b","Type":"ContainerStarted","Data":"84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5"} Nov 24 07:56:13 crc kubenswrapper[4799]: I1124 07:56:13.154702 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4tfxf" podStartSLOduration=2.65725823 podStartE2EDuration="5.154682943s" podCreationTimestamp="2025-11-24 07:56:08 +0000 UTC" firstStartedPulling="2025-11-24 07:56:10.076123266 +0000 UTC m=+4115.732105740" lastFinishedPulling="2025-11-24 07:56:12.573547979 +0000 UTC m=+4118.229530453" observedRunningTime="2025-11-24 07:56:13.149439314 +0000 UTC m=+4118.805421878" watchObservedRunningTime="2025-11-24 07:56:13.154682943 +0000 UTC m=+4118.810665407" Nov 24 07:56:18 crc kubenswrapper[4799]: I1124 07:56:18.944838 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:18 crc kubenswrapper[4799]: I1124 07:56:18.945642 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:20 crc kubenswrapper[4799]: I1124 07:56:20.012068 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4tfxf" podUID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerName="registry-server" probeResult="failure" output=< Nov 24 07:56:20 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 07:56:20 crc kubenswrapper[4799]: > Nov 24 07:56:26 crc kubenswrapper[4799]: I1124 07:56:26.894600 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d5bf6b659-rvhmt"] Nov 24 07:56:26 crc kubenswrapper[4799]: I1124 07:56:26.896092 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:26 crc kubenswrapper[4799]: I1124 07:56:26.897953 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 07:56:26 crc kubenswrapper[4799]: I1124 07:56:26.897974 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-rh9hh" Nov 24 07:56:26 crc kubenswrapper[4799]: I1124 07:56:26.898150 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 07:56:26 crc kubenswrapper[4799]: I1124 07:56:26.898281 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 07:56:26 crc kubenswrapper[4799]: I1124 07:56:26.899549 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 07:56:26 crc kubenswrapper[4799]: I1124 07:56:26.908575 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d5bf6b659-rvhmt"] Nov 24 07:56:26 crc kubenswrapper[4799]: I1124 07:56:26.925466 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-dns-svc\") pod \"dnsmasq-dns-7d5bf6b659-rvhmt\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:26 crc kubenswrapper[4799]: I1124 07:56:26.925711 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-config\") pod \"dnsmasq-dns-7d5bf6b659-rvhmt\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:26 crc kubenswrapper[4799]: I1124 07:56:26.925766 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdcfs\" (UniqueName: \"kubernetes.io/projected/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-kube-api-access-qdcfs\") pod \"dnsmasq-dns-7d5bf6b659-rvhmt\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.026653 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-dns-svc\") pod \"dnsmasq-dns-7d5bf6b659-rvhmt\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.026713 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-config\") pod \"dnsmasq-dns-7d5bf6b659-rvhmt\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.026783 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdcfs\" (UniqueName: \"kubernetes.io/projected/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-kube-api-access-qdcfs\") pod \"dnsmasq-dns-7d5bf6b659-rvhmt\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.028070 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-dns-svc\") pod \"dnsmasq-dns-7d5bf6b659-rvhmt\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.028765 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-config\") pod \"dnsmasq-dns-7d5bf6b659-rvhmt\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.049661 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdcfs\" (UniqueName: \"kubernetes.io/projected/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-kube-api-access-qdcfs\") pod \"dnsmasq-dns-7d5bf6b659-rvhmt\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.159106 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b76c5459c-4gx87"] Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.160501 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.169872 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b76c5459c-4gx87"] Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.214021 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.228545 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-dns-svc\") pod \"dnsmasq-dns-7b76c5459c-4gx87\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.228958 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-config\") pod \"dnsmasq-dns-7b76c5459c-4gx87\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.229023 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfxlc\" (UniqueName: \"kubernetes.io/projected/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-kube-api-access-cfxlc\") pod \"dnsmasq-dns-7b76c5459c-4gx87\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.330001 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfxlc\" (UniqueName: \"kubernetes.io/projected/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-kube-api-access-cfxlc\") pod \"dnsmasq-dns-7b76c5459c-4gx87\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.330061 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-dns-svc\") pod \"dnsmasq-dns-7b76c5459c-4gx87\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.330116 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-config\") pod \"dnsmasq-dns-7b76c5459c-4gx87\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.330935 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-config\") pod \"dnsmasq-dns-7b76c5459c-4gx87\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.331061 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-dns-svc\") pod \"dnsmasq-dns-7b76c5459c-4gx87\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.375779 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfxlc\" (UniqueName: \"kubernetes.io/projected/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-kube-api-access-cfxlc\") pod \"dnsmasq-dns-7b76c5459c-4gx87\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.481282 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:27 crc kubenswrapper[4799]: W1124 07:56:27.784354 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f3ed7c8_a7d4_4489_b6cf_051a00edc1df.slice/crio-eb5ac0af1a54619c769f57ecaee1b6df7cdda6c92bf177a82bfc7cc63329d893 WatchSource:0}: Error finding container eb5ac0af1a54619c769f57ecaee1b6df7cdda6c92bf177a82bfc7cc63329d893: Status 404 returned error can't find the container with id eb5ac0af1a54619c769f57ecaee1b6df7cdda6c92bf177a82bfc7cc63329d893 Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.789155 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d5bf6b659-rvhmt"] Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.936020 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.936898 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.938485 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-8vj7l" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.938985 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 07:56:27 crc kubenswrapper[4799]: I1124 07:56:27.945150 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.039915 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h7xv\" (UniqueName: \"kubernetes.io/projected/e99ee713-7523-4dde-9751-34c9ef7bad1f-kube-api-access-4h7xv\") pod \"memcached-0\" (UID: \"e99ee713-7523-4dde-9751-34c9ef7bad1f\") " pod="openstack/memcached-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.039987 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e99ee713-7523-4dde-9751-34c9ef7bad1f-kolla-config\") pod \"memcached-0\" (UID: \"e99ee713-7523-4dde-9751-34c9ef7bad1f\") " pod="openstack/memcached-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.040192 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e99ee713-7523-4dde-9751-34c9ef7bad1f-config-data\") pod \"memcached-0\" (UID: \"e99ee713-7523-4dde-9751-34c9ef7bad1f\") " pod="openstack/memcached-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.050831 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b76c5459c-4gx87"] Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.086587 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.090046 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.092507 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.095841 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.096064 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.096179 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.096296 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-sl242" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.096418 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.141823 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e99ee713-7523-4dde-9751-34c9ef7bad1f-config-data\") pod \"memcached-0\" (UID: \"e99ee713-7523-4dde-9751-34c9ef7bad1f\") " pod="openstack/memcached-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.141908 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h7xv\" (UniqueName: \"kubernetes.io/projected/e99ee713-7523-4dde-9751-34c9ef7bad1f-kube-api-access-4h7xv\") pod \"memcached-0\" (UID: \"e99ee713-7523-4dde-9751-34c9ef7bad1f\") " pod="openstack/memcached-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.141951 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e99ee713-7523-4dde-9751-34c9ef7bad1f-kolla-config\") pod \"memcached-0\" (UID: \"e99ee713-7523-4dde-9751-34c9ef7bad1f\") " pod="openstack/memcached-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.142731 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e99ee713-7523-4dde-9751-34c9ef7bad1f-kolla-config\") pod \"memcached-0\" (UID: \"e99ee713-7523-4dde-9751-34c9ef7bad1f\") " pod="openstack/memcached-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.144245 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e99ee713-7523-4dde-9751-34c9ef7bad1f-config-data\") pod \"memcached-0\" (UID: \"e99ee713-7523-4dde-9751-34c9ef7bad1f\") " pod="openstack/memcached-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.181672 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h7xv\" (UniqueName: \"kubernetes.io/projected/e99ee713-7523-4dde-9751-34c9ef7bad1f-kube-api-access-4h7xv\") pod \"memcached-0\" (UID: \"e99ee713-7523-4dde-9751-34c9ef7bad1f\") " pod="openstack/memcached-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.244995 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.245077 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trgv6\" (UniqueName: \"kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-kube-api-access-trgv6\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.245141 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.245169 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.245217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.245375 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.245536 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.245680 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.245703 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.255616 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.292086 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" event={"ID":"85c5f2a8-b835-4efe-9c2e-42deb39aeb66","Type":"ContainerStarted","Data":"21641aef784a32c684b11f050dc262e8e8ce0222d957b6f7742372393173c400"} Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.294412 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" event={"ID":"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df","Type":"ContainerStarted","Data":"eb5ac0af1a54619c769f57ecaee1b6df7cdda6c92bf177a82bfc7cc63329d893"} Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.305880 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.307250 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.314326 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.315779 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.316147 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-8jj5f" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.316564 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.320677 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.323196 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.346500 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.346586 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.346621 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.346658 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.346692 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trgv6\" (UniqueName: \"kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-kube-api-access-trgv6\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.346735 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.346767 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.346796 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.346826 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.348063 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.348952 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.349391 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.351265 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.351941 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.352355 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.368830 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.368919 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c41590325b7cf08665d36acaf946212d686a1b6bc959beff8a6186cdcb0c9a83/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.371875 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.372751 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trgv6\" (UniqueName: \"kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-kube-api-access-trgv6\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.374405 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.377040 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.378504 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.378682 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.378923 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hhg9b" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.379097 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.379186 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.385006 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.411258 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\") pod \"rabbitmq-server-0\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.427151 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.449289 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-600a350d-230d-4b9b-8523-576bb69c0a56\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-600a350d-230d-4b9b-8523-576bb69c0a56\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.449349 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/63e8e2e8-1c56-48b1-a0fe-5626a960209f-config-data-default\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.449367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/63e8e2e8-1c56-48b1-a0fe-5626a960209f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.449397 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63e8e2e8-1c56-48b1-a0fe-5626a960209f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.449424 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/63e8e2e8-1c56-48b1-a0fe-5626a960209f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.449440 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9trck\" (UniqueName: \"kubernetes.io/projected/63e8e2e8-1c56-48b1-a0fe-5626a960209f-kube-api-access-9trck\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.449468 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/63e8e2e8-1c56-48b1-a0fe-5626a960209f-kolla-config\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.449487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e8e2e8-1c56-48b1-a0fe-5626a960209f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550191 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rksck\" (UniqueName: \"kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-kube-api-access-rksck\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550252 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-600a350d-230d-4b9b-8523-576bb69c0a56\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-600a350d-230d-4b9b-8523-576bb69c0a56\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550274 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550307 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/63e8e2e8-1c56-48b1-a0fe-5626a960209f-config-data-default\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550325 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/63e8e2e8-1c56-48b1-a0fe-5626a960209f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550353 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63e8e2e8-1c56-48b1-a0fe-5626a960209f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550375 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550390 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/63e8e2e8-1c56-48b1-a0fe-5626a960209f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550405 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9trck\" (UniqueName: \"kubernetes.io/projected/63e8e2e8-1c56-48b1-a0fe-5626a960209f-kube-api-access-9trck\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550424 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550439 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550466 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/63e8e2e8-1c56-48b1-a0fe-5626a960209f-kolla-config\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550481 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550499 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e8e2e8-1c56-48b1-a0fe-5626a960209f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550516 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-93b2e989-96b0-4682-9232-0949eacc676a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550538 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.550555 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.555904 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.555938 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-600a350d-230d-4b9b-8523-576bb69c0a56\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-600a350d-230d-4b9b-8523-576bb69c0a56\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/98ec2523081ae576e6be9c1d0a709e55ba00d5eb5ec6c378cf626b55f90ca7f2/globalmount\"" pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.556462 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/63e8e2e8-1c56-48b1-a0fe-5626a960209f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.556706 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e8e2e8-1c56-48b1-a0fe-5626a960209f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.556954 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/63e8e2e8-1c56-48b1-a0fe-5626a960209f-config-data-default\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.557124 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/63e8e2e8-1c56-48b1-a0fe-5626a960209f-kolla-config\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.559654 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/63e8e2e8-1c56-48b1-a0fe-5626a960209f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.559683 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63e8e2e8-1c56-48b1-a0fe-5626a960209f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.566925 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9trck\" (UniqueName: \"kubernetes.io/projected/63e8e2e8-1c56-48b1-a0fe-5626a960209f-kube-api-access-9trck\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.593665 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-600a350d-230d-4b9b-8523-576bb69c0a56\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-600a350d-230d-4b9b-8523-576bb69c0a56\") pod \"openstack-galera-0\" (UID: \"63e8e2e8-1c56-48b1-a0fe-5626a960209f\") " pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.648986 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.651488 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.651534 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-93b2e989-96b0-4682-9232-0949eacc676a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.651564 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.651581 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.651607 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rksck\" (UniqueName: \"kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-kube-api-access-rksck\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.651644 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.651697 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.651719 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.651737 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.652656 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.655383 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.655656 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-93b2e989-96b0-4682-9232-0949eacc676a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ba9e29719594b0b6510f2334d36bf34bab0509a397ecf78c0711574db5f987e2/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.655392 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.655790 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.656264 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.658015 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.659046 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.662799 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.668332 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rksck\" (UniqueName: \"kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-kube-api-access-rksck\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.690412 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-93b2e989-96b0-4682-9232-0949eacc676a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a\") pod \"rabbitmq-cell1-server-0\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.706707 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.790368 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 07:56:28 crc kubenswrapper[4799]: I1124 07:56:28.921146 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.065213 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.128789 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.135535 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.135637 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.140237 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.141161 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-m4jdg" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.141481 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.141644 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.157076 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:56:29 crc kubenswrapper[4799]: W1124 07:56:29.178410 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0bc6d2f0_b01c_4007_a1d7_f21e3eba48e2.slice/crio-0851864651f1416bf351c29c50c86065032309499f8bfb245b99b863a37718c6 WatchSource:0}: Error finding container 0851864651f1416bf351c29c50c86065032309499f8bfb245b99b863a37718c6: Status 404 returned error can't find the container with id 0851864651f1416bf351c29c50c86065032309499f8bfb245b99b863a37718c6 Nov 24 07:56:29 crc kubenswrapper[4799]: W1124 07:56:29.182244 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8e2e8_1c56_48b1_a0fe_5626a960209f.slice/crio-584efe2370c584ac36dfe6661b75c74c1ff80ea840e86160b77e077a3631c37e WatchSource:0}: Error finding container 584efe2370c584ac36dfe6661b75c74c1ff80ea840e86160b77e077a3631c37e: Status 404 returned error can't find the container with id 584efe2370c584ac36dfe6661b75c74c1ff80ea840e86160b77e077a3631c37e Nov 24 07:56:29 crc kubenswrapper[4799]: W1124 07:56:29.184661 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6959053_4c2e_4f43_b11d_12ec72e3f2bd.slice/crio-4fae8d7a246c9d1a80aec3c42ed7604f29d75e77d20dcdd292fd511efe184f8f WatchSource:0}: Error finding container 4fae8d7a246c9d1a80aec3c42ed7604f29d75e77d20dcdd292fd511efe184f8f: Status 404 returned error can't find the container with id 4fae8d7a246c9d1a80aec3c42ed7604f29d75e77d20dcdd292fd511efe184f8f Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.261422 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7kzw\" (UniqueName: \"kubernetes.io/projected/2d77e145-8599-45f3-ab56-13e97d34183e-kube-api-access-s7kzw\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.261463 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d77e145-8599-45f3-ab56-13e97d34183e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.261495 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d77e145-8599-45f3-ab56-13e97d34183e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.261553 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-562b33c0-7b55-4850-8866-a14e0db70b48\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-562b33c0-7b55-4850-8866-a14e0db70b48\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.261573 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d77e145-8599-45f3-ab56-13e97d34183e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.261606 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d77e145-8599-45f3-ab56-13e97d34183e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.261637 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d77e145-8599-45f3-ab56-13e97d34183e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.261659 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d77e145-8599-45f3-ab56-13e97d34183e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.302863 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2","Type":"ContainerStarted","Data":"0851864651f1416bf351c29c50c86065032309499f8bfb245b99b863a37718c6"} Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.304668 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"63e8e2e8-1c56-48b1-a0fe-5626a960209f","Type":"ContainerStarted","Data":"584efe2370c584ac36dfe6661b75c74c1ff80ea840e86160b77e077a3631c37e"} Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.305707 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e99ee713-7523-4dde-9751-34c9ef7bad1f","Type":"ContainerStarted","Data":"cafac0bfc8b43fa368b9d8eae021817805e0f79b6748d1bcfe11415ada0541dc"} Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.309532 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b6959053-4c2e-4f43-b11d-12ec72e3f2bd","Type":"ContainerStarted","Data":"4fae8d7a246c9d1a80aec3c42ed7604f29d75e77d20dcdd292fd511efe184f8f"} Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.313547 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.360576 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.362522 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7kzw\" (UniqueName: \"kubernetes.io/projected/2d77e145-8599-45f3-ab56-13e97d34183e-kube-api-access-s7kzw\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.362562 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d77e145-8599-45f3-ab56-13e97d34183e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.362593 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d77e145-8599-45f3-ab56-13e97d34183e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.362644 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-562b33c0-7b55-4850-8866-a14e0db70b48\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-562b33c0-7b55-4850-8866-a14e0db70b48\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.362661 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d77e145-8599-45f3-ab56-13e97d34183e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.362687 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d77e145-8599-45f3-ab56-13e97d34183e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.362710 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d77e145-8599-45f3-ab56-13e97d34183e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.362732 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d77e145-8599-45f3-ab56-13e97d34183e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.363306 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2d77e145-8599-45f3-ab56-13e97d34183e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.363702 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2d77e145-8599-45f3-ab56-13e97d34183e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.363713 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2d77e145-8599-45f3-ab56-13e97d34183e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.364511 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d77e145-8599-45f3-ab56-13e97d34183e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.365442 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.365481 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-562b33c0-7b55-4850-8866-a14e0db70b48\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-562b33c0-7b55-4850-8866-a14e0db70b48\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1e6440322d5c985191979a3bcf116dec281f1c80cebd130af212771dca6c640d/globalmount\"" pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.367745 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d77e145-8599-45f3-ab56-13e97d34183e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.367749 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d77e145-8599-45f3-ab56-13e97d34183e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.387321 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7kzw\" (UniqueName: \"kubernetes.io/projected/2d77e145-8599-45f3-ab56-13e97d34183e-kube-api-access-s7kzw\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.402078 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-562b33c0-7b55-4850-8866-a14e0db70b48\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-562b33c0-7b55-4850-8866-a14e0db70b48\") pod \"openstack-cell1-galera-0\" (UID: \"2d77e145-8599-45f3-ab56-13e97d34183e\") " pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.542070 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4tfxf"] Nov 24 07:56:29 crc kubenswrapper[4799]: I1124 07:56:29.662744 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.268410 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 07:56:31 crc kubenswrapper[4799]: W1124 07:56:31.276449 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d77e145_8599_45f3_ab56_13e97d34183e.slice/crio-439eef693b4ff57c93140dd354f3967245805b90249e87961b23db9029982088 WatchSource:0}: Error finding container 439eef693b4ff57c93140dd354f3967245805b90249e87961b23db9029982088: Status 404 returned error can't find the container with id 439eef693b4ff57c93140dd354f3967245805b90249e87961b23db9029982088 Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.323955 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d77e145-8599-45f3-ab56-13e97d34183e","Type":"ContainerStarted","Data":"439eef693b4ff57c93140dd354f3967245805b90249e87961b23db9029982088"} Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.324092 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4tfxf" podUID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerName="registry-server" containerID="cri-o://84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5" gracePeriod=2 Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.722064 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.798529 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2v8p\" (UniqueName: \"kubernetes.io/projected/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-kube-api-access-d2v8p\") pod \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.798700 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-catalog-content\") pod \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.798737 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-utilities\") pod \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\" (UID: \"14a602c5-0fe1-4b5c-b15c-6fc532fd863b\") " Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.799796 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-utilities" (OuterVolumeSpecName: "utilities") pod "14a602c5-0fe1-4b5c-b15c-6fc532fd863b" (UID: "14a602c5-0fe1-4b5c-b15c-6fc532fd863b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.804778 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-kube-api-access-d2v8p" (OuterVolumeSpecName: "kube-api-access-d2v8p") pod "14a602c5-0fe1-4b5c-b15c-6fc532fd863b" (UID: "14a602c5-0fe1-4b5c-b15c-6fc532fd863b"). InnerVolumeSpecName "kube-api-access-d2v8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.900525 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2v8p\" (UniqueName: \"kubernetes.io/projected/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-kube-api-access-d2v8p\") on node \"crc\" DevicePath \"\"" Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.900586 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:56:31 crc kubenswrapper[4799]: I1124 07:56:31.912613 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14a602c5-0fe1-4b5c-b15c-6fc532fd863b" (UID: "14a602c5-0fe1-4b5c-b15c-6fc532fd863b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.002356 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a602c5-0fe1-4b5c-b15c-6fc532fd863b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.336197 4799 generic.go:334] "Generic (PLEG): container finished" podID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerID="84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5" exitCode=0 Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.336251 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4tfxf" Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.336246 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4tfxf" event={"ID":"14a602c5-0fe1-4b5c-b15c-6fc532fd863b","Type":"ContainerDied","Data":"84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5"} Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.336356 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4tfxf" event={"ID":"14a602c5-0fe1-4b5c-b15c-6fc532fd863b","Type":"ContainerDied","Data":"bb6bbfb15105837e265b55b66fb93ee8eb86999a05a47f515734f4c7aef25790"} Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.336381 4799 scope.go:117] "RemoveContainer" containerID="84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5" Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.373583 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4tfxf"] Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.377792 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4tfxf"] Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.393173 4799 scope.go:117] "RemoveContainer" containerID="8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854" Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.411451 4799 scope.go:117] "RemoveContainer" containerID="0ca7495bbf2c43a9e6db5ad564e87edea3e0851ff413435d7a94b178213cd397" Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.450603 4799 scope.go:117] "RemoveContainer" containerID="84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5" Nov 24 07:56:32 crc kubenswrapper[4799]: E1124 07:56:32.451518 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5\": container with ID starting with 84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5 not found: ID does not exist" containerID="84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5" Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.451563 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5"} err="failed to get container status \"84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5\": rpc error: code = NotFound desc = could not find container \"84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5\": container with ID starting with 84786572df4298b11904f570b27f2571e9ae012437332ae9396b25b1e3d9e8d5 not found: ID does not exist" Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.451596 4799 scope.go:117] "RemoveContainer" containerID="8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854" Nov 24 07:56:32 crc kubenswrapper[4799]: E1124 07:56:32.451825 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854\": container with ID starting with 8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854 not found: ID does not exist" containerID="8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854" Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.451876 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854"} err="failed to get container status \"8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854\": rpc error: code = NotFound desc = could not find container \"8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854\": container with ID starting with 8467f55ee908dc162d6054dd6b2749281ab37b17323feb72956316cc55cde854 not found: ID does not exist" Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.451891 4799 scope.go:117] "RemoveContainer" containerID="0ca7495bbf2c43a9e6db5ad564e87edea3e0851ff413435d7a94b178213cd397" Nov 24 07:56:32 crc kubenswrapper[4799]: E1124 07:56:32.452554 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ca7495bbf2c43a9e6db5ad564e87edea3e0851ff413435d7a94b178213cd397\": container with ID starting with 0ca7495bbf2c43a9e6db5ad564e87edea3e0851ff413435d7a94b178213cd397 not found: ID does not exist" containerID="0ca7495bbf2c43a9e6db5ad564e87edea3e0851ff413435d7a94b178213cd397" Nov 24 07:56:32 crc kubenswrapper[4799]: I1124 07:56:32.452576 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca7495bbf2c43a9e6db5ad564e87edea3e0851ff413435d7a94b178213cd397"} err="failed to get container status \"0ca7495bbf2c43a9e6db5ad564e87edea3e0851ff413435d7a94b178213cd397\": rpc error: code = NotFound desc = could not find container \"0ca7495bbf2c43a9e6db5ad564e87edea3e0851ff413435d7a94b178213cd397\": container with ID starting with 0ca7495bbf2c43a9e6db5ad564e87edea3e0851ff413435d7a94b178213cd397 not found: ID does not exist" Nov 24 07:56:33 crc kubenswrapper[4799]: I1124 07:56:33.641838 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" path="/var/lib/kubelet/pods/14a602c5-0fe1-4b5c-b15c-6fc532fd863b/volumes" Nov 24 07:56:44 crc kubenswrapper[4799]: I1124 07:56:44.445288 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e99ee713-7523-4dde-9751-34c9ef7bad1f","Type":"ContainerStarted","Data":"0493c5fc3c0612e070831c4ab9ccb51b52d266c4c8951c99afc27e8bac0d9ad2"} Nov 24 07:56:44 crc kubenswrapper[4799]: I1124 07:56:44.445953 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 07:56:44 crc kubenswrapper[4799]: I1124 07:56:44.469979 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.72233129 podStartE2EDuration="17.46995203s" podCreationTimestamp="2025-11-24 07:56:27 +0000 UTC" firstStartedPulling="2025-11-24 07:56:29.178832143 +0000 UTC m=+4134.834814617" lastFinishedPulling="2025-11-24 07:56:43.926452873 +0000 UTC m=+4149.582435357" observedRunningTime="2025-11-24 07:56:44.462222731 +0000 UTC m=+4150.118205215" watchObservedRunningTime="2025-11-24 07:56:44.46995203 +0000 UTC m=+4150.125934544" Nov 24 07:56:45 crc kubenswrapper[4799]: I1124 07:56:45.454539 4799 generic.go:334] "Generic (PLEG): container finished" podID="2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" containerID="86208b05f7c719c6bd7b86330797626f7f9bb76b2c496fa481e81e2fc523f9a2" exitCode=0 Nov 24 07:56:45 crc kubenswrapper[4799]: I1124 07:56:45.454616 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" event={"ID":"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df","Type":"ContainerDied","Data":"86208b05f7c719c6bd7b86330797626f7f9bb76b2c496fa481e81e2fc523f9a2"} Nov 24 07:56:45 crc kubenswrapper[4799]: I1124 07:56:45.457296 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d77e145-8599-45f3-ab56-13e97d34183e","Type":"ContainerStarted","Data":"a06fe8ed7b7357abb9bedbd18de2bf2ff7e7f280eb7bc357bf1f4556f0e6fa1e"} Nov 24 07:56:45 crc kubenswrapper[4799]: I1124 07:56:45.461423 4799 generic.go:334] "Generic (PLEG): container finished" podID="85c5f2a8-b835-4efe-9c2e-42deb39aeb66" containerID="620939a3b8502244a00e4498fb771081bad6e9ef94e2e1b763815fe9011c74c4" exitCode=0 Nov 24 07:56:45 crc kubenswrapper[4799]: I1124 07:56:45.461505 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" event={"ID":"85c5f2a8-b835-4efe-9c2e-42deb39aeb66","Type":"ContainerDied","Data":"620939a3b8502244a00e4498fb771081bad6e9ef94e2e1b763815fe9011c74c4"} Nov 24 07:56:45 crc kubenswrapper[4799]: I1124 07:56:45.463537 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"63e8e2e8-1c56-48b1-a0fe-5626a960209f","Type":"ContainerStarted","Data":"2870964ce341199ac03709a35114cb494c7eb2bb186d1bd9366e130bf4bbdcb8"} Nov 24 07:56:45 crc kubenswrapper[4799]: E1124 07:56:45.571969 4799 mount_linux.go:282] Mount failed: exit status 32 Nov 24 07:56:45 crc kubenswrapper[4799]: Mounting command: mount Nov 24 07:56:45 crc kubenswrapper[4799]: Mounting arguments: --no-canonicalize -o bind /proc/4799/fd/26 /var/lib/kubelet/pods/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df/volume-subpaths/dns-svc/dnsmasq-dns/1 Nov 24 07:56:45 crc kubenswrapper[4799]: Output: mount: /var/lib/kubelet/pods/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df/volume-subpaths/dns-svc/dnsmasq-dns/1: mount(2) system call failed: No such file or directory. Nov 24 07:56:45 crc kubenswrapper[4799]: E1124 07:56:45.598459 4799 kubelet_pods.go:349] "Failed to prepare subPath for volumeMount of the container" err=< Nov 24 07:56:45 crc kubenswrapper[4799]: error mounting /var/lib/kubelet/pods/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df/volumes/kubernetes.io~configmap/dns-svc/..2025_11_24_07_56_27.2198309859/dns-svc: mount failed: exit status 32 Nov 24 07:56:45 crc kubenswrapper[4799]: Mounting command: mount Nov 24 07:56:45 crc kubenswrapper[4799]: Mounting arguments: --no-canonicalize -o bind /proc/4799/fd/26 /var/lib/kubelet/pods/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df/volume-subpaths/dns-svc/dnsmasq-dns/1 Nov 24 07:56:45 crc kubenswrapper[4799]: Output: mount: /var/lib/kubelet/pods/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df/volume-subpaths/dns-svc/dnsmasq-dns/1: mount(2) system call failed: No such file or directory. Nov 24 07:56:45 crc kubenswrapper[4799]: > containerName="dnsmasq-dns" volumeMountName="dns-svc" Nov 24 07:56:45 crc kubenswrapper[4799]: E1124 07:56:45.598649 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:dnsmasq-dns,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:94a34fda2d142cfe9e3097b1d1bd6839,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8chc6h5bh56fh546hb7hc8h67h5bchffh577h697h5b5h5bdh59bhf6hf4h558hb5h578h595h5cchfbh644h59ch7fh654h547h587h5cbh5d5h8fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qdcfs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7d5bf6b659-rvhmt_openstack(2f3ed7c8-a7d4-4489-b6cf-051a00edc1df): CreateContainerConfigError: failed to prepare subPath for volumeMount \"dns-svc\" of container \"dnsmasq-dns\"" logger="UnhandledError" Nov 24 07:56:45 crc kubenswrapper[4799]: E1124 07:56:45.600591 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerConfigError: \"failed to prepare subPath for volumeMount \\\"dns-svc\\\" of container \\\"dnsmasq-dns\\\"\"" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" podUID="2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" Nov 24 07:56:46 crc kubenswrapper[4799]: I1124 07:56:46.475567 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b6959053-4c2e-4f43-b11d-12ec72e3f2bd","Type":"ContainerStarted","Data":"ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a"} Nov 24 07:56:46 crc kubenswrapper[4799]: I1124 07:56:46.478089 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2","Type":"ContainerStarted","Data":"0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede"} Nov 24 07:56:46 crc kubenswrapper[4799]: I1124 07:56:46.481804 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" event={"ID":"85c5f2a8-b835-4efe-9c2e-42deb39aeb66","Type":"ContainerStarted","Data":"0d07ae3e5dbfd1104e91b3b2573a1ea98b8dd2e6a847756746ca87f013111e34"} Nov 24 07:56:46 crc kubenswrapper[4799]: I1124 07:56:46.482177 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:46 crc kubenswrapper[4799]: I1124 07:56:46.535039 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" podStartSLOduration=3.638193879 podStartE2EDuration="19.535014947s" podCreationTimestamp="2025-11-24 07:56:27 +0000 UTC" firstStartedPulling="2025-11-24 07:56:28.098730055 +0000 UTC m=+4133.754712529" lastFinishedPulling="2025-11-24 07:56:43.995551123 +0000 UTC m=+4149.651533597" observedRunningTime="2025-11-24 07:56:46.532693111 +0000 UTC m=+4152.188675625" watchObservedRunningTime="2025-11-24 07:56:46.535014947 +0000 UTC m=+4152.190997441" Nov 24 07:56:47 crc kubenswrapper[4799]: I1124 07:56:47.490791 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" event={"ID":"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df","Type":"ContainerStarted","Data":"567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032"} Nov 24 07:56:47 crc kubenswrapper[4799]: I1124 07:56:47.514689 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" podStartSLOduration=5.336593879 podStartE2EDuration="21.514668326s" podCreationTimestamp="2025-11-24 07:56:26 +0000 UTC" firstStartedPulling="2025-11-24 07:56:27.786521648 +0000 UTC m=+4133.442504122" lastFinishedPulling="2025-11-24 07:56:43.964596085 +0000 UTC m=+4149.620578569" observedRunningTime="2025-11-24 07:56:47.508732718 +0000 UTC m=+4153.164715202" watchObservedRunningTime="2025-11-24 07:56:47.514668326 +0000 UTC m=+4153.170650810" Nov 24 07:56:48 crc kubenswrapper[4799]: I1124 07:56:48.503668 4799 generic.go:334] "Generic (PLEG): container finished" podID="63e8e2e8-1c56-48b1-a0fe-5626a960209f" containerID="2870964ce341199ac03709a35114cb494c7eb2bb186d1bd9366e130bf4bbdcb8" exitCode=0 Nov 24 07:56:48 crc kubenswrapper[4799]: I1124 07:56:48.503843 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"63e8e2e8-1c56-48b1-a0fe-5626a960209f","Type":"ContainerDied","Data":"2870964ce341199ac03709a35114cb494c7eb2bb186d1bd9366e130bf4bbdcb8"} Nov 24 07:56:48 crc kubenswrapper[4799]: I1124 07:56:48.510758 4799 generic.go:334] "Generic (PLEG): container finished" podID="2d77e145-8599-45f3-ab56-13e97d34183e" containerID="a06fe8ed7b7357abb9bedbd18de2bf2ff7e7f280eb7bc357bf1f4556f0e6fa1e" exitCode=0 Nov 24 07:56:48 crc kubenswrapper[4799]: I1124 07:56:48.510825 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d77e145-8599-45f3-ab56-13e97d34183e","Type":"ContainerDied","Data":"a06fe8ed7b7357abb9bedbd18de2bf2ff7e7f280eb7bc357bf1f4556f0e6fa1e"} Nov 24 07:56:49 crc kubenswrapper[4799]: I1124 07:56:49.528568 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"63e8e2e8-1c56-48b1-a0fe-5626a960209f","Type":"ContainerStarted","Data":"1cebb93c97a4953fa6b31d01441a36e9de9826518a79a82770028fadff481aa7"} Nov 24 07:56:49 crc kubenswrapper[4799]: I1124 07:56:49.532078 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2d77e145-8599-45f3-ab56-13e97d34183e","Type":"ContainerStarted","Data":"31e96379af7975a6f402772e14e3ffad26e551ed27543fce97e3fbf35b61bacf"} Nov 24 07:56:49 crc kubenswrapper[4799]: I1124 07:56:49.596950 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.836683744 podStartE2EDuration="22.596917061s" podCreationTimestamp="2025-11-24 07:56:27 +0000 UTC" firstStartedPulling="2025-11-24 07:56:29.191507433 +0000 UTC m=+4134.847489907" lastFinishedPulling="2025-11-24 07:56:43.95174072 +0000 UTC m=+4149.607723224" observedRunningTime="2025-11-24 07:56:49.558354957 +0000 UTC m=+4155.214337511" watchObservedRunningTime="2025-11-24 07:56:49.596917061 +0000 UTC m=+4155.252899605" Nov 24 07:56:49 crc kubenswrapper[4799]: I1124 07:56:49.600377 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=8.915784819 podStartE2EDuration="21.600366339s" podCreationTimestamp="2025-11-24 07:56:28 +0000 UTC" firstStartedPulling="2025-11-24 07:56:31.280981422 +0000 UTC m=+4136.936963896" lastFinishedPulling="2025-11-24 07:56:43.965562912 +0000 UTC m=+4149.621545416" observedRunningTime="2025-11-24 07:56:49.593439712 +0000 UTC m=+4155.249422216" watchObservedRunningTime="2025-11-24 07:56:49.600366339 +0000 UTC m=+4155.256348853" Nov 24 07:56:49 crc kubenswrapper[4799]: I1124 07:56:49.663215 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:49 crc kubenswrapper[4799]: I1124 07:56:49.663277 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:51 crc kubenswrapper[4799]: E1124 07:56:51.683065 4799 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:39144->38.102.83.50:45263: write tcp 38.102.83.50:39144->38.102.83.50:45263: write: broken pipe Nov 24 07:56:52 crc kubenswrapper[4799]: I1124 07:56:52.214659 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:52 crc kubenswrapper[4799]: I1124 07:56:52.216067 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:52 crc kubenswrapper[4799]: I1124 07:56:52.482994 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:56:52 crc kubenswrapper[4799]: I1124 07:56:52.529261 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d5bf6b659-rvhmt"] Nov 24 07:56:53 crc kubenswrapper[4799]: I1124 07:56:53.257307 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 07:56:53 crc kubenswrapper[4799]: I1124 07:56:53.568317 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" podUID="2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" containerName="dnsmasq-dns" containerID="cri-o://567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032" gracePeriod=10 Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.008632 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.095690 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-dns-svc\") pod \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.096361 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-config\") pod \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.096401 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdcfs\" (UniqueName: \"kubernetes.io/projected/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-kube-api-access-qdcfs\") pod \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\" (UID: \"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df\") " Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.101422 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-kube-api-access-qdcfs" (OuterVolumeSpecName: "kube-api-access-qdcfs") pod "2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" (UID: "2f3ed7c8-a7d4-4489-b6cf-051a00edc1df"). InnerVolumeSpecName "kube-api-access-qdcfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.132113 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" (UID: "2f3ed7c8-a7d4-4489-b6cf-051a00edc1df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.132300 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-config" (OuterVolumeSpecName: "config") pod "2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" (UID: "2f3ed7c8-a7d4-4489-b6cf-051a00edc1df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.198005 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.198038 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdcfs\" (UniqueName: \"kubernetes.io/projected/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-kube-api-access-qdcfs\") on node \"crc\" DevicePath \"\"" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.198051 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.581644 4799 generic.go:334] "Generic (PLEG): container finished" podID="2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" containerID="567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032" exitCode=0 Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.581693 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" event={"ID":"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df","Type":"ContainerDied","Data":"567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032"} Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.581724 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" event={"ID":"2f3ed7c8-a7d4-4489-b6cf-051a00edc1df","Type":"ContainerDied","Data":"eb5ac0af1a54619c769f57ecaee1b6df7cdda6c92bf177a82bfc7cc63329d893"} Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.581741 4799 scope.go:117] "RemoveContainer" containerID="567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.581816 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5bf6b659-rvhmt" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.607343 4799 scope.go:117] "RemoveContainer" containerID="86208b05f7c719c6bd7b86330797626f7f9bb76b2c496fa481e81e2fc523f9a2" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.627803 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d5bf6b659-rvhmt"] Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.634259 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d5bf6b659-rvhmt"] Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.658904 4799 scope.go:117] "RemoveContainer" containerID="567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032" Nov 24 07:56:54 crc kubenswrapper[4799]: E1124 07:56:54.659485 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032\": container with ID starting with 567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032 not found: ID does not exist" containerID="567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.659541 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032"} err="failed to get container status \"567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032\": rpc error: code = NotFound desc = could not find container \"567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032\": container with ID starting with 567f5fb8de8b36cac3adac79ba036f1a4884348b8bdc5062bd849531586d4032 not found: ID does not exist" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.659572 4799 scope.go:117] "RemoveContainer" containerID="86208b05f7c719c6bd7b86330797626f7f9bb76b2c496fa481e81e2fc523f9a2" Nov 24 07:56:54 crc kubenswrapper[4799]: E1124 07:56:54.660141 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86208b05f7c719c6bd7b86330797626f7f9bb76b2c496fa481e81e2fc523f9a2\": container with ID starting with 86208b05f7c719c6bd7b86330797626f7f9bb76b2c496fa481e81e2fc523f9a2 not found: ID does not exist" containerID="86208b05f7c719c6bd7b86330797626f7f9bb76b2c496fa481e81e2fc523f9a2" Nov 24 07:56:54 crc kubenswrapper[4799]: I1124 07:56:54.660195 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86208b05f7c719c6bd7b86330797626f7f9bb76b2c496fa481e81e2fc523f9a2"} err="failed to get container status \"86208b05f7c719c6bd7b86330797626f7f9bb76b2c496fa481e81e2fc523f9a2\": rpc error: code = NotFound desc = could not find container \"86208b05f7c719c6bd7b86330797626f7f9bb76b2c496fa481e81e2fc523f9a2\": container with ID starting with 86208b05f7c719c6bd7b86330797626f7f9bb76b2c496fa481e81e2fc523f9a2 not found: ID does not exist" Nov 24 07:56:55 crc kubenswrapper[4799]: I1124 07:56:55.640204 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" path="/var/lib/kubelet/pods/2f3ed7c8-a7d4-4489-b6cf-051a00edc1df/volumes" Nov 24 07:56:55 crc kubenswrapper[4799]: I1124 07:56:55.783162 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:55 crc kubenswrapper[4799]: I1124 07:56:55.890989 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 07:56:58 crc kubenswrapper[4799]: I1124 07:56:58.650567 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 07:56:58 crc kubenswrapper[4799]: I1124 07:56:58.650902 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 07:56:58 crc kubenswrapper[4799]: I1124 07:56:58.744492 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 07:56:59 crc kubenswrapper[4799]: I1124 07:56:59.714897 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 07:57:18 crc kubenswrapper[4799]: I1124 07:57:18.804525 4799 generic.go:334] "Generic (PLEG): container finished" podID="b6959053-4c2e-4f43-b11d-12ec72e3f2bd" containerID="ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a" exitCode=0 Nov 24 07:57:18 crc kubenswrapper[4799]: I1124 07:57:18.804658 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b6959053-4c2e-4f43-b11d-12ec72e3f2bd","Type":"ContainerDied","Data":"ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a"} Nov 24 07:57:18 crc kubenswrapper[4799]: I1124 07:57:18.807688 4799 generic.go:334] "Generic (PLEG): container finished" podID="0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" containerID="0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede" exitCode=0 Nov 24 07:57:18 crc kubenswrapper[4799]: I1124 07:57:18.807753 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2","Type":"ContainerDied","Data":"0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede"} Nov 24 07:57:19 crc kubenswrapper[4799]: I1124 07:57:19.823813 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2","Type":"ContainerStarted","Data":"4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027"} Nov 24 07:57:19 crc kubenswrapper[4799]: I1124 07:57:19.824504 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 07:57:19 crc kubenswrapper[4799]: I1124 07:57:19.826587 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b6959053-4c2e-4f43-b11d-12ec72e3f2bd","Type":"ContainerStarted","Data":"e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122"} Nov 24 07:57:19 crc kubenswrapper[4799]: I1124 07:57:19.827283 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:19 crc kubenswrapper[4799]: I1124 07:57:19.848530 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.075543147 podStartE2EDuration="52.848506486s" podCreationTimestamp="2025-11-24 07:56:27 +0000 UTC" firstStartedPulling="2025-11-24 07:56:29.182675852 +0000 UTC m=+4134.838658326" lastFinishedPulling="2025-11-24 07:56:43.955639181 +0000 UTC m=+4149.611621665" observedRunningTime="2025-11-24 07:57:19.84370986 +0000 UTC m=+4185.499692354" watchObservedRunningTime="2025-11-24 07:57:19.848506486 +0000 UTC m=+4185.504488960" Nov 24 07:57:19 crc kubenswrapper[4799]: I1124 07:57:19.870002 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.160062674 podStartE2EDuration="52.869980885s" podCreationTimestamp="2025-11-24 07:56:27 +0000 UTC" firstStartedPulling="2025-11-24 07:56:29.187012765 +0000 UTC m=+4134.842995239" lastFinishedPulling="2025-11-24 07:56:43.896930966 +0000 UTC m=+4149.552913450" observedRunningTime="2025-11-24 07:57:19.867286639 +0000 UTC m=+4185.523269163" watchObservedRunningTime="2025-11-24 07:57:19.869980885 +0000 UTC m=+4185.525963359" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.685120 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hqvzb"] Nov 24 07:57:28 crc kubenswrapper[4799]: E1124 07:57:28.686427 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerName="registry-server" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.686455 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerName="registry-server" Nov 24 07:57:28 crc kubenswrapper[4799]: E1124 07:57:28.686494 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerName="extract-utilities" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.686507 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerName="extract-utilities" Nov 24 07:57:28 crc kubenswrapper[4799]: E1124 07:57:28.686532 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" containerName="init" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.686546 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" containerName="init" Nov 24 07:57:28 crc kubenswrapper[4799]: E1124 07:57:28.686574 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" containerName="dnsmasq-dns" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.686587 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" containerName="dnsmasq-dns" Nov 24 07:57:28 crc kubenswrapper[4799]: E1124 07:57:28.686618 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerName="extract-content" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.686630 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerName="extract-content" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.686959 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a602c5-0fe1-4b5c-b15c-6fc532fd863b" containerName="registry-server" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.686984 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3ed7c8-a7d4-4489-b6cf-051a00edc1df" containerName="dnsmasq-dns" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.690704 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.701203 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hqvzb"] Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.787155 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb5s2\" (UniqueName: \"kubernetes.io/projected/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-kube-api-access-cb5s2\") pod \"certified-operators-hqvzb\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.787209 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-catalog-content\") pod \"certified-operators-hqvzb\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.787238 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-utilities\") pod \"certified-operators-hqvzb\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.888332 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb5s2\" (UniqueName: \"kubernetes.io/projected/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-kube-api-access-cb5s2\") pod \"certified-operators-hqvzb\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.888382 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-catalog-content\") pod \"certified-operators-hqvzb\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.888409 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-utilities\") pod \"certified-operators-hqvzb\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.888987 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-utilities\") pod \"certified-operators-hqvzb\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.889128 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-catalog-content\") pod \"certified-operators-hqvzb\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:28 crc kubenswrapper[4799]: I1124 07:57:28.913553 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb5s2\" (UniqueName: \"kubernetes.io/projected/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-kube-api-access-cb5s2\") pod \"certified-operators-hqvzb\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:29 crc kubenswrapper[4799]: I1124 07:57:29.058600 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:29 crc kubenswrapper[4799]: I1124 07:57:29.562062 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hqvzb"] Nov 24 07:57:29 crc kubenswrapper[4799]: W1124 07:57:29.571028 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea54c1e3_0e38_41f1_95c7_996fe0e52d5d.slice/crio-9477a5198875b351d5bc9ee3a55a3b957b7fbf0a8b63026ab84166d0b19040e8 WatchSource:0}: Error finding container 9477a5198875b351d5bc9ee3a55a3b957b7fbf0a8b63026ab84166d0b19040e8: Status 404 returned error can't find the container with id 9477a5198875b351d5bc9ee3a55a3b957b7fbf0a8b63026ab84166d0b19040e8 Nov 24 07:57:29 crc kubenswrapper[4799]: I1124 07:57:29.903113 4799 generic.go:334] "Generic (PLEG): container finished" podID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" containerID="eca3d381366079e3bf953313004ec8652678ac40be1a931f23766a1dafef080d" exitCode=0 Nov 24 07:57:29 crc kubenswrapper[4799]: I1124 07:57:29.903155 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqvzb" event={"ID":"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d","Type":"ContainerDied","Data":"eca3d381366079e3bf953313004ec8652678ac40be1a931f23766a1dafef080d"} Nov 24 07:57:29 crc kubenswrapper[4799]: I1124 07:57:29.903185 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqvzb" event={"ID":"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d","Type":"ContainerStarted","Data":"9477a5198875b351d5bc9ee3a55a3b957b7fbf0a8b63026ab84166d0b19040e8"} Nov 24 07:57:29 crc kubenswrapper[4799]: I1124 07:57:29.904703 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 07:57:30 crc kubenswrapper[4799]: I1124 07:57:30.918087 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqvzb" event={"ID":"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d","Type":"ContainerStarted","Data":"323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5"} Nov 24 07:57:31 crc kubenswrapper[4799]: I1124 07:57:31.935474 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqvzb" event={"ID":"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d","Type":"ContainerDied","Data":"323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5"} Nov 24 07:57:31 crc kubenswrapper[4799]: I1124 07:57:31.936479 4799 generic.go:334] "Generic (PLEG): container finished" podID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" containerID="323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5" exitCode=0 Nov 24 07:57:32 crc kubenswrapper[4799]: I1124 07:57:32.953780 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqvzb" event={"ID":"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d","Type":"ContainerStarted","Data":"2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483"} Nov 24 07:57:32 crc kubenswrapper[4799]: I1124 07:57:32.983704 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hqvzb" podStartSLOduration=2.29641083 podStartE2EDuration="4.983683697s" podCreationTimestamp="2025-11-24 07:57:28 +0000 UTC" firstStartedPulling="2025-11-24 07:57:29.904275737 +0000 UTC m=+4195.560258251" lastFinishedPulling="2025-11-24 07:57:32.591548604 +0000 UTC m=+4198.247531118" observedRunningTime="2025-11-24 07:57:32.978967873 +0000 UTC m=+4198.634950387" watchObservedRunningTime="2025-11-24 07:57:32.983683697 +0000 UTC m=+4198.639666181" Nov 24 07:57:38 crc kubenswrapper[4799]: I1124 07:57:38.913451 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:38 crc kubenswrapper[4799]: I1124 07:57:38.915526 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 07:57:39 crc kubenswrapper[4799]: I1124 07:57:39.059012 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:39 crc kubenswrapper[4799]: I1124 07:57:39.059089 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:39 crc kubenswrapper[4799]: I1124 07:57:39.121178 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:40 crc kubenswrapper[4799]: I1124 07:57:40.101309 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:41 crc kubenswrapper[4799]: I1124 07:57:41.073968 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hqvzb"] Nov 24 07:57:42 crc kubenswrapper[4799]: I1124 07:57:42.039629 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hqvzb" podUID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" containerName="registry-server" containerID="cri-o://2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483" gracePeriod=2 Nov 24 07:57:42 crc kubenswrapper[4799]: I1124 07:57:42.425066 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:42 crc kubenswrapper[4799]: I1124 07:57:42.523338 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb5s2\" (UniqueName: \"kubernetes.io/projected/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-kube-api-access-cb5s2\") pod \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " Nov 24 07:57:42 crc kubenswrapper[4799]: I1124 07:57:42.523440 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-utilities\") pod \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " Nov 24 07:57:42 crc kubenswrapper[4799]: I1124 07:57:42.523465 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-catalog-content\") pod \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\" (UID: \"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d\") " Nov 24 07:57:42 crc kubenswrapper[4799]: I1124 07:57:42.525678 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-utilities" (OuterVolumeSpecName: "utilities") pod "ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" (UID: "ea54c1e3-0e38-41f1-95c7-996fe0e52d5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:57:42 crc kubenswrapper[4799]: I1124 07:57:42.529224 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-kube-api-access-cb5s2" (OuterVolumeSpecName: "kube-api-access-cb5s2") pod "ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" (UID: "ea54c1e3-0e38-41f1-95c7-996fe0e52d5d"). InnerVolumeSpecName "kube-api-access-cb5s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:57:42 crc kubenswrapper[4799]: I1124 07:57:42.577433 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" (UID: "ea54c1e3-0e38-41f1-95c7-996fe0e52d5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:57:42 crc kubenswrapper[4799]: I1124 07:57:42.625604 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb5s2\" (UniqueName: \"kubernetes.io/projected/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-kube-api-access-cb5s2\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:42 crc kubenswrapper[4799]: I1124 07:57:42.625662 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:42 crc kubenswrapper[4799]: I1124 07:57:42.625678 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.054025 4799 generic.go:334] "Generic (PLEG): container finished" podID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" containerID="2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483" exitCode=0 Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.054093 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqvzb" event={"ID":"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d","Type":"ContainerDied","Data":"2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483"} Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.054565 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqvzb" event={"ID":"ea54c1e3-0e38-41f1-95c7-996fe0e52d5d","Type":"ContainerDied","Data":"9477a5198875b351d5bc9ee3a55a3b957b7fbf0a8b63026ab84166d0b19040e8"} Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.054646 4799 scope.go:117] "RemoveContainer" containerID="2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.054113 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqvzb" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.089303 4799 scope.go:117] "RemoveContainer" containerID="323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.120185 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hqvzb"] Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.129938 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hqvzb"] Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.130084 4799 scope.go:117] "RemoveContainer" containerID="eca3d381366079e3bf953313004ec8652678ac40be1a931f23766a1dafef080d" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.161261 4799 scope.go:117] "RemoveContainer" containerID="2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483" Nov 24 07:57:43 crc kubenswrapper[4799]: E1124 07:57:43.161708 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483\": container with ID starting with 2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483 not found: ID does not exist" containerID="2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.161742 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483"} err="failed to get container status \"2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483\": rpc error: code = NotFound desc = could not find container \"2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483\": container with ID starting with 2738cf32fdf6f1368b5797adfb46b9f01ca4c531867a32f8a6cc4d3543c38483 not found: ID does not exist" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.161761 4799 scope.go:117] "RemoveContainer" containerID="323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5" Nov 24 07:57:43 crc kubenswrapper[4799]: E1124 07:57:43.162304 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5\": container with ID starting with 323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5 not found: ID does not exist" containerID="323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.162330 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5"} err="failed to get container status \"323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5\": rpc error: code = NotFound desc = could not find container \"323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5\": container with ID starting with 323966e0b6422aaeb2efcd2635e0c31fad97ddf4cc1741c7f0b7fa21a9e955b5 not found: ID does not exist" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.162351 4799 scope.go:117] "RemoveContainer" containerID="eca3d381366079e3bf953313004ec8652678ac40be1a931f23766a1dafef080d" Nov 24 07:57:43 crc kubenswrapper[4799]: E1124 07:57:43.162619 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eca3d381366079e3bf953313004ec8652678ac40be1a931f23766a1dafef080d\": container with ID starting with eca3d381366079e3bf953313004ec8652678ac40be1a931f23766a1dafef080d not found: ID does not exist" containerID="eca3d381366079e3bf953313004ec8652678ac40be1a931f23766a1dafef080d" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.162668 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca3d381366079e3bf953313004ec8652678ac40be1a931f23766a1dafef080d"} err="failed to get container status \"eca3d381366079e3bf953313004ec8652678ac40be1a931f23766a1dafef080d\": rpc error: code = NotFound desc = could not find container \"eca3d381366079e3bf953313004ec8652678ac40be1a931f23766a1dafef080d\": container with ID starting with eca3d381366079e3bf953313004ec8652678ac40be1a931f23766a1dafef080d not found: ID does not exist" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.488189 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b997d949-97n87"] Nov 24 07:57:43 crc kubenswrapper[4799]: E1124 07:57:43.488569 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" containerName="extract-utilities" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.488581 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" containerName="extract-utilities" Nov 24 07:57:43 crc kubenswrapper[4799]: E1124 07:57:43.488607 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" containerName="extract-content" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.488613 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" containerName="extract-content" Nov 24 07:57:43 crc kubenswrapper[4799]: E1124 07:57:43.488629 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" containerName="registry-server" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.488636 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" containerName="registry-server" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.488784 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" containerName="registry-server" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.490294 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.506821 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b997d949-97n87"] Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.540920 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-config\") pod \"dnsmasq-dns-5b997d949-97n87\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.541071 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv2dd\" (UniqueName: \"kubernetes.io/projected/26016233-fb28-44e0-8490-6af2dc9bcb62-kube-api-access-bv2dd\") pod \"dnsmasq-dns-5b997d949-97n87\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.541124 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-dns-svc\") pod \"dnsmasq-dns-5b997d949-97n87\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.640710 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea54c1e3-0e38-41f1-95c7-996fe0e52d5d" path="/var/lib/kubelet/pods/ea54c1e3-0e38-41f1-95c7-996fe0e52d5d/volumes" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.642130 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv2dd\" (UniqueName: \"kubernetes.io/projected/26016233-fb28-44e0-8490-6af2dc9bcb62-kube-api-access-bv2dd\") pod \"dnsmasq-dns-5b997d949-97n87\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.642196 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-dns-svc\") pod \"dnsmasq-dns-5b997d949-97n87\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.642249 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-config\") pod \"dnsmasq-dns-5b997d949-97n87\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.643459 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-config\") pod \"dnsmasq-dns-5b997d949-97n87\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.643715 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-dns-svc\") pod \"dnsmasq-dns-5b997d949-97n87\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.667847 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv2dd\" (UniqueName: \"kubernetes.io/projected/26016233-fb28-44e0-8490-6af2dc9bcb62-kube-api-access-bv2dd\") pod \"dnsmasq-dns-5b997d949-97n87\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:43 crc kubenswrapper[4799]: I1124 07:57:43.810600 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:44 crc kubenswrapper[4799]: I1124 07:57:44.200229 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:57:44 crc kubenswrapper[4799]: I1124 07:57:44.267981 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b997d949-97n87"] Nov 24 07:57:44 crc kubenswrapper[4799]: I1124 07:57:44.840283 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:57:45 crc kubenswrapper[4799]: I1124 07:57:45.071634 4799 generic.go:334] "Generic (PLEG): container finished" podID="26016233-fb28-44e0-8490-6af2dc9bcb62" containerID="33e59c740ed24dfff89969e0dff5aa08aae7d761799ec19347e190f722c68f60" exitCode=0 Nov 24 07:57:45 crc kubenswrapper[4799]: I1124 07:57:45.071689 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b997d949-97n87" event={"ID":"26016233-fb28-44e0-8490-6af2dc9bcb62","Type":"ContainerDied","Data":"33e59c740ed24dfff89969e0dff5aa08aae7d761799ec19347e190f722c68f60"} Nov 24 07:57:45 crc kubenswrapper[4799]: I1124 07:57:45.071749 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b997d949-97n87" event={"ID":"26016233-fb28-44e0-8490-6af2dc9bcb62","Type":"ContainerStarted","Data":"8b341fbf9c10fc4718986fe8472de09618f4c94012ca28065b390708282d1e2e"} Nov 24 07:57:46 crc kubenswrapper[4799]: I1124 07:57:46.079590 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b997d949-97n87" event={"ID":"26016233-fb28-44e0-8490-6af2dc9bcb62","Type":"ContainerStarted","Data":"7b7632164ae632dfc4974107fb150033d94f21134ecfe906e4f5bdd4ad38af0a"} Nov 24 07:57:46 crc kubenswrapper[4799]: I1124 07:57:46.079886 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:46 crc kubenswrapper[4799]: I1124 07:57:46.089390 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" containerName="rabbitmq" containerID="cri-o://4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027" gracePeriod=604799 Nov 24 07:57:46 crc kubenswrapper[4799]: I1124 07:57:46.097677 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b997d949-97n87" podStartSLOduration=3.097655121 podStartE2EDuration="3.097655121s" podCreationTimestamp="2025-11-24 07:57:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:57:46.094754888 +0000 UTC m=+4211.750737362" watchObservedRunningTime="2025-11-24 07:57:46.097655121 +0000 UTC m=+4211.753637595" Nov 24 07:57:46 crc kubenswrapper[4799]: I1124 07:57:46.584463 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="b6959053-4c2e-4f43-b11d-12ec72e3f2bd" containerName="rabbitmq" containerID="cri-o://e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122" gracePeriod=604799 Nov 24 07:57:48 crc kubenswrapper[4799]: I1124 07:57:48.428165 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.239:5672: connect: connection refused" Nov 24 07:57:48 crc kubenswrapper[4799]: I1124 07:57:48.707370 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b6959053-4c2e-4f43-b11d-12ec72e3f2bd" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.241:5672: connect: connection refused" Nov 24 07:57:50 crc kubenswrapper[4799]: I1124 07:57:50.400658 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:57:50 crc kubenswrapper[4799]: I1124 07:57:50.402475 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.665085 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.808023 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-pod-info\") pod \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.808116 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-plugins-conf\") pod \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.808192 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-server-conf\") pod \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.808222 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-erlang-cookie-secret\") pod \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.808357 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\") pod \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.808400 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-plugins\") pod \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.808432 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-confd\") pod \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.808470 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-erlang-cookie\") pod \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.808506 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trgv6\" (UniqueName: \"kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-kube-api-access-trgv6\") pod \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\" (UID: \"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2\") " Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.808782 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" (UID: "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.809213 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" (UID: "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.809968 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" (UID: "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.825242 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-pod-info" (OuterVolumeSpecName: "pod-info") pod "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" (UID: "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.825326 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" (UID: "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.825795 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109" (OuterVolumeSpecName: "persistence") pod "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" (UID: "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2"). InnerVolumeSpecName "pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.828287 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-kube-api-access-trgv6" (OuterVolumeSpecName: "kube-api-access-trgv6") pod "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" (UID: "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2"). InnerVolumeSpecName "kube-api-access-trgv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.830218 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-server-conf" (OuterVolumeSpecName: "server-conf") pod "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" (UID: "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.910275 4799 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.910305 4799 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.910317 4799 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.910328 4799 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.910364 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\") on node \"crc\" " Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.910378 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.910390 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.910403 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trgv6\" (UniqueName: \"kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-kube-api-access-trgv6\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.925511 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" (UID: "0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.926991 4799 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 07:57:52 crc kubenswrapper[4799]: I1124 07:57:52.927144 4799 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109") on node "crc" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.011478 4799 reconciler_common.go:293] "Volume detached for volume \"pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.011511 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.118954 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.149574 4799 generic.go:334] "Generic (PLEG): container finished" podID="0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" containerID="4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027" exitCode=0 Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.149640 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2","Type":"ContainerDied","Data":"4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027"} Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.149682 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2","Type":"ContainerDied","Data":"0851864651f1416bf351c29c50c86065032309499f8bfb245b99b863a37718c6"} Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.149702 4799 scope.go:117] "RemoveContainer" containerID="4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.149785 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.159111 4799 generic.go:334] "Generic (PLEG): container finished" podID="b6959053-4c2e-4f43-b11d-12ec72e3f2bd" containerID="e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122" exitCode=0 Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.159151 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b6959053-4c2e-4f43-b11d-12ec72e3f2bd","Type":"ContainerDied","Data":"e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122"} Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.159200 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b6959053-4c2e-4f43-b11d-12ec72e3f2bd","Type":"ContainerDied","Data":"4fae8d7a246c9d1a80aec3c42ed7604f29d75e77d20dcdd292fd511efe184f8f"} Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.159266 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.180197 4799 scope.go:117] "RemoveContainer" containerID="0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.197958 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.209138 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.215862 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-erlang-cookie-secret\") pod \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.215889 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-plugins-conf\") pod \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.215907 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rksck\" (UniqueName: \"kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-kube-api-access-rksck\") pod \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.215937 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-server-conf\") pod \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.215979 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-erlang-cookie\") pod \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.216051 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-pod-info\") pod \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.216146 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a\") pod \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.216169 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-confd\") pod \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.216218 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-plugins\") pod \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\" (UID: \"b6959053-4c2e-4f43-b11d-12ec72e3f2bd\") " Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.216790 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b6959053-4c2e-4f43-b11d-12ec72e3f2bd" (UID: "b6959053-4c2e-4f43-b11d-12ec72e3f2bd"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.219279 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b6959053-4c2e-4f43-b11d-12ec72e3f2bd" (UID: "b6959053-4c2e-4f43-b11d-12ec72e3f2bd"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.219586 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b6959053-4c2e-4f43-b11d-12ec72e3f2bd" (UID: "b6959053-4c2e-4f43-b11d-12ec72e3f2bd"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.228131 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-pod-info" (OuterVolumeSpecName: "pod-info") pod "b6959053-4c2e-4f43-b11d-12ec72e3f2bd" (UID: "b6959053-4c2e-4f43-b11d-12ec72e3f2bd"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.228604 4799 scope.go:117] "RemoveContainer" containerID="4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.228631 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b6959053-4c2e-4f43-b11d-12ec72e3f2bd" (UID: "b6959053-4c2e-4f43-b11d-12ec72e3f2bd"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.228829 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-kube-api-access-rksck" (OuterVolumeSpecName: "kube-api-access-rksck") pod "b6959053-4c2e-4f43-b11d-12ec72e3f2bd" (UID: "b6959053-4c2e-4f43-b11d-12ec72e3f2bd"). InnerVolumeSpecName "kube-api-access-rksck". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:57:53 crc kubenswrapper[4799]: E1124 07:57:53.233336 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027\": container with ID starting with 4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027 not found: ID does not exist" containerID="4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.233370 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027"} err="failed to get container status \"4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027\": rpc error: code = NotFound desc = could not find container \"4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027\": container with ID starting with 4acf6c1205855bdaba4bfd2e3f8fc4a6f797713c0d22d87857ed8f8d3dc5e027 not found: ID does not exist" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.233395 4799 scope.go:117] "RemoveContainer" containerID="0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede" Nov 24 07:57:53 crc kubenswrapper[4799]: E1124 07:57:53.233696 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede\": container with ID starting with 0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede not found: ID does not exist" containerID="0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.233720 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede"} err="failed to get container status \"0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede\": rpc error: code = NotFound desc = could not find container \"0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede\": container with ID starting with 0fba994514ad667fc250c6b64b42f36bff5e03432d2050ccd3b4231f60152ede not found: ID does not exist" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.233734 4799 scope.go:117] "RemoveContainer" containerID="e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.242955 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:57:53 crc kubenswrapper[4799]: E1124 07:57:53.243702 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6959053-4c2e-4f43-b11d-12ec72e3f2bd" containerName="rabbitmq" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.243724 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6959053-4c2e-4f43-b11d-12ec72e3f2bd" containerName="rabbitmq" Nov 24 07:57:53 crc kubenswrapper[4799]: E1124 07:57:53.243744 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" containerName="setup-container" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.243753 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" containerName="setup-container" Nov 24 07:57:53 crc kubenswrapper[4799]: E1124 07:57:53.243776 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6959053-4c2e-4f43-b11d-12ec72e3f2bd" containerName="setup-container" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.243784 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6959053-4c2e-4f43-b11d-12ec72e3f2bd" containerName="setup-container" Nov 24 07:57:53 crc kubenswrapper[4799]: E1124 07:57:53.243797 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" containerName="rabbitmq" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.243805 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" containerName="rabbitmq" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.244084 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6959053-4c2e-4f43-b11d-12ec72e3f2bd" containerName="rabbitmq" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.244112 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" containerName="rabbitmq" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.245124 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.247060 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.249224 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.249258 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-sl242" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.249306 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.249557 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.249673 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a" (OuterVolumeSpecName: "persistence") pod "b6959053-4c2e-4f43-b11d-12ec72e3f2bd" (UID: "b6959053-4c2e-4f43-b11d-12ec72e3f2bd"). InnerVolumeSpecName "pvc-93b2e989-96b0-4682-9232-0949eacc676a". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.249715 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.258157 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-server-conf" (OuterVolumeSpecName: "server-conf") pod "b6959053-4c2e-4f43-b11d-12ec72e3f2bd" (UID: "b6959053-4c2e-4f43-b11d-12ec72e3f2bd"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.264013 4799 scope.go:117] "RemoveContainer" containerID="ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.288822 4799 scope.go:117] "RemoveContainer" containerID="e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122" Nov 24 07:57:53 crc kubenswrapper[4799]: E1124 07:57:53.289568 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122\": container with ID starting with e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122 not found: ID does not exist" containerID="e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.289625 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122"} err="failed to get container status \"e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122\": rpc error: code = NotFound desc = could not find container \"e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122\": container with ID starting with e273bd8935af2182704eb56cbe85b2ead2273f4cb66c5b302a34e025a04de122 not found: ID does not exist" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.289669 4799 scope.go:117] "RemoveContainer" containerID="ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a" Nov 24 07:57:53 crc kubenswrapper[4799]: E1124 07:57:53.289988 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a\": container with ID starting with ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a not found: ID does not exist" containerID="ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.290023 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a"} err="failed to get container status \"ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a\": rpc error: code = NotFound desc = could not find container \"ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a\": container with ID starting with ab617d9632bd088b0eb9a473f2fc25cd803726be1c8cbe2708e96844eadaae8a not found: ID does not exist" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318307 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2bde1fe-98c3-4302-a350-72cfbc703dd2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318381 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2bde1fe-98c3-4302-a350-72cfbc703dd2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318408 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7bvq\" (UniqueName: \"kubernetes.io/projected/f2bde1fe-98c3-4302-a350-72cfbc703dd2-kube-api-access-j7bvq\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318450 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2bde1fe-98c3-4302-a350-72cfbc703dd2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318466 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2bde1fe-98c3-4302-a350-72cfbc703dd2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2bde1fe-98c3-4302-a350-72cfbc703dd2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318517 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2bde1fe-98c3-4302-a350-72cfbc703dd2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318551 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318571 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2bde1fe-98c3-4302-a350-72cfbc703dd2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318611 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318621 4799 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318629 4799 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318638 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rksck\" (UniqueName: \"kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-kube-api-access-rksck\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318647 4799 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318656 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318665 4799 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.318687 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-93b2e989-96b0-4682-9232-0949eacc676a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a\") on node \"crc\" " Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.333094 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b6959053-4c2e-4f43-b11d-12ec72e3f2bd" (UID: "b6959053-4c2e-4f43-b11d-12ec72e3f2bd"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.335890 4799 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.336021 4799 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-93b2e989-96b0-4682-9232-0949eacc676a" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a") on node "crc" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.419382 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2bde1fe-98c3-4302-a350-72cfbc703dd2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.419443 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.419505 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2bde1fe-98c3-4302-a350-72cfbc703dd2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.420149 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2bde1fe-98c3-4302-a350-72cfbc703dd2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.420525 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2bde1fe-98c3-4302-a350-72cfbc703dd2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.420935 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2bde1fe-98c3-4302-a350-72cfbc703dd2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.420982 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2bde1fe-98c3-4302-a350-72cfbc703dd2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.421015 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7bvq\" (UniqueName: \"kubernetes.io/projected/f2bde1fe-98c3-4302-a350-72cfbc703dd2-kube-api-access-j7bvq\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.421757 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2bde1fe-98c3-4302-a350-72cfbc703dd2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.421780 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2bde1fe-98c3-4302-a350-72cfbc703dd2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.421804 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2bde1fe-98c3-4302-a350-72cfbc703dd2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.421889 4799 reconciler_common.go:293] "Volume detached for volume \"pvc-93b2e989-96b0-4682-9232-0949eacc676a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.421908 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b6959053-4c2e-4f43-b11d-12ec72e3f2bd-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.422517 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2bde1fe-98c3-4302-a350-72cfbc703dd2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.423822 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.423885 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c41590325b7cf08665d36acaf946212d686a1b6bc959beff8a6186cdcb0c9a83/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.424072 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2bde1fe-98c3-4302-a350-72cfbc703dd2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.424166 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2bde1fe-98c3-4302-a350-72cfbc703dd2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.424396 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2bde1fe-98c3-4302-a350-72cfbc703dd2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.425954 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2bde1fe-98c3-4302-a350-72cfbc703dd2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.445703 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7bvq\" (UniqueName: \"kubernetes.io/projected/f2bde1fe-98c3-4302-a350-72cfbc703dd2-kube-api-access-j7bvq\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.454488 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f43afd8a-cc39-4d2f-8a9c-f7253ee5c109\") pod \"rabbitmq-server-0\" (UID: \"f2bde1fe-98c3-4302-a350-72cfbc703dd2\") " pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.529197 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.531184 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.558678 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.564022 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.566382 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hhg9b" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.566588 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.566808 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.567026 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.567640 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.570506 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.577351 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.625601 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.625663 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggtcc\" (UniqueName: \"kubernetes.io/projected/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-kube-api-access-ggtcc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.625716 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.625743 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.625769 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.625814 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.625918 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-93b2e989-96b0-4682-9232-0949eacc676a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.626660 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.632195 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.644771 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2" path="/var/lib/kubelet/pods/0bc6d2f0-b01c-4007-a1d7-f21e3eba48e2/volumes" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.645518 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6959053-4c2e-4f43-b11d-12ec72e3f2bd" path="/var/lib/kubelet/pods/b6959053-4c2e-4f43-b11d-12ec72e3f2bd/volumes" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.734493 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.734568 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggtcc\" (UniqueName: \"kubernetes.io/projected/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-kube-api-access-ggtcc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.734590 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.734612 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.734639 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.734684 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.734746 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-93b2e989-96b0-4682-9232-0949eacc676a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.734770 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.734798 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.735557 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.735612 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.735975 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.736146 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.737418 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.737443 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-93b2e989-96b0-4682-9232-0949eacc676a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ba9e29719594b0b6510f2334d36bf34bab0509a397ecf78c0711574db5f987e2/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.740937 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.741285 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.741650 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.751256 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggtcc\" (UniqueName: \"kubernetes.io/projected/2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4-kube-api-access-ggtcc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.775078 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-93b2e989-96b0-4682-9232-0949eacc676a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b2e989-96b0-4682-9232-0949eacc676a\") pod \"rabbitmq-cell1-server-0\" (UID: \"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.790962 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.811024 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.878254 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b76c5459c-4gx87"] Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.878526 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" podUID="85c5f2a8-b835-4efe-9c2e-42deb39aeb66" containerName="dnsmasq-dns" containerID="cri-o://0d07ae3e5dbfd1104e91b3b2573a1ea98b8dd2e6a847756746ca87f013111e34" gracePeriod=10 Nov 24 07:57:53 crc kubenswrapper[4799]: I1124 07:57:53.930777 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.167551 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f2bde1fe-98c3-4302-a350-72cfbc703dd2","Type":"ContainerStarted","Data":"1bbd9a44834fbf1a94a3dd8a2e8ff67e215c7e573ec49831bf3724571c8587f1"} Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.170416 4799 generic.go:334] "Generic (PLEG): container finished" podID="85c5f2a8-b835-4efe-9c2e-42deb39aeb66" containerID="0d07ae3e5dbfd1104e91b3b2573a1ea98b8dd2e6a847756746ca87f013111e34" exitCode=0 Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.170449 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" event={"ID":"85c5f2a8-b835-4efe-9c2e-42deb39aeb66","Type":"ContainerDied","Data":"0d07ae3e5dbfd1104e91b3b2573a1ea98b8dd2e6a847756746ca87f013111e34"} Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.249343 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.346004 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-config\") pod \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.346074 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfxlc\" (UniqueName: \"kubernetes.io/projected/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-kube-api-access-cfxlc\") pod \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.346113 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-dns-svc\") pod \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\" (UID: \"85c5f2a8-b835-4efe-9c2e-42deb39aeb66\") " Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.354602 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-kube-api-access-cfxlc" (OuterVolumeSpecName: "kube-api-access-cfxlc") pod "85c5f2a8-b835-4efe-9c2e-42deb39aeb66" (UID: "85c5f2a8-b835-4efe-9c2e-42deb39aeb66"). InnerVolumeSpecName "kube-api-access-cfxlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.399678 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-config" (OuterVolumeSpecName: "config") pod "85c5f2a8-b835-4efe-9c2e-42deb39aeb66" (UID: "85c5f2a8-b835-4efe-9c2e-42deb39aeb66"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.447189 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-config\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.447223 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfxlc\" (UniqueName: \"kubernetes.io/projected/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-kube-api-access-cfxlc\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.476264 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 07:57:54 crc kubenswrapper[4799]: W1124 07:57:54.478080 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a41cd9d_3e31_4ff5_be57_1cc067c9c0f4.slice/crio-02ec32a6ae9a31bffa7813145d172304b0f759d89fb3aff5f6e3f67ddb03f554 WatchSource:0}: Error finding container 02ec32a6ae9a31bffa7813145d172304b0f759d89fb3aff5f6e3f67ddb03f554: Status 404 returned error can't find the container with id 02ec32a6ae9a31bffa7813145d172304b0f759d89fb3aff5f6e3f67ddb03f554 Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.502922 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "85c5f2a8-b835-4efe-9c2e-42deb39aeb66" (UID: "85c5f2a8-b835-4efe-9c2e-42deb39aeb66"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 07:57:54 crc kubenswrapper[4799]: I1124 07:57:54.550270 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85c5f2a8-b835-4efe-9c2e-42deb39aeb66-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 07:57:55 crc kubenswrapper[4799]: I1124 07:57:55.183064 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4","Type":"ContainerStarted","Data":"02ec32a6ae9a31bffa7813145d172304b0f759d89fb3aff5f6e3f67ddb03f554"} Nov 24 07:57:55 crc kubenswrapper[4799]: I1124 07:57:55.186914 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" event={"ID":"85c5f2a8-b835-4efe-9c2e-42deb39aeb66","Type":"ContainerDied","Data":"21641aef784a32c684b11f050dc262e8e8ce0222d957b6f7742372393173c400"} Nov 24 07:57:55 crc kubenswrapper[4799]: I1124 07:57:55.186963 4799 scope.go:117] "RemoveContainer" containerID="0d07ae3e5dbfd1104e91b3b2573a1ea98b8dd2e6a847756746ca87f013111e34" Nov 24 07:57:55 crc kubenswrapper[4799]: I1124 07:57:55.186988 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b76c5459c-4gx87" Nov 24 07:57:55 crc kubenswrapper[4799]: I1124 07:57:55.189896 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f2bde1fe-98c3-4302-a350-72cfbc703dd2","Type":"ContainerStarted","Data":"b786a90713afdd1e8396846c7bd7b309d794d27082f21ab6fbd7aeeac68c9eeb"} Nov 24 07:57:55 crc kubenswrapper[4799]: I1124 07:57:55.216147 4799 scope.go:117] "RemoveContainer" containerID="620939a3b8502244a00e4498fb771081bad6e9ef94e2e1b763815fe9011c74c4" Nov 24 07:57:55 crc kubenswrapper[4799]: I1124 07:57:55.274475 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b76c5459c-4gx87"] Nov 24 07:57:55 crc kubenswrapper[4799]: I1124 07:57:55.281314 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b76c5459c-4gx87"] Nov 24 07:57:55 crc kubenswrapper[4799]: I1124 07:57:55.638844 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85c5f2a8-b835-4efe-9c2e-42deb39aeb66" path="/var/lib/kubelet/pods/85c5f2a8-b835-4efe-9c2e-42deb39aeb66/volumes" Nov 24 07:57:56 crc kubenswrapper[4799]: I1124 07:57:56.197914 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4","Type":"ContainerStarted","Data":"5cfa07f9c8617bb6e62c98dde6c86aa231ff8c9b60dd667879d2890f7b33086b"} Nov 24 07:58:20 crc kubenswrapper[4799]: I1124 07:58:20.401091 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:58:20 crc kubenswrapper[4799]: I1124 07:58:20.402074 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:58:28 crc kubenswrapper[4799]: I1124 07:58:28.537763 4799 generic.go:334] "Generic (PLEG): container finished" podID="f2bde1fe-98c3-4302-a350-72cfbc703dd2" containerID="b786a90713afdd1e8396846c7bd7b309d794d27082f21ab6fbd7aeeac68c9eeb" exitCode=0 Nov 24 07:58:28 crc kubenswrapper[4799]: I1124 07:58:28.537896 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f2bde1fe-98c3-4302-a350-72cfbc703dd2","Type":"ContainerDied","Data":"b786a90713afdd1e8396846c7bd7b309d794d27082f21ab6fbd7aeeac68c9eeb"} Nov 24 07:58:29 crc kubenswrapper[4799]: I1124 07:58:29.553312 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f2bde1fe-98c3-4302-a350-72cfbc703dd2","Type":"ContainerStarted","Data":"61dbe558b926e00dcbfde3bba688c936ca43f9167a0b403d3727e8812a62fd27"} Nov 24 07:58:29 crc kubenswrapper[4799]: I1124 07:58:29.553939 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 07:58:29 crc kubenswrapper[4799]: I1124 07:58:29.557225 4799 generic.go:334] "Generic (PLEG): container finished" podID="2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4" containerID="5cfa07f9c8617bb6e62c98dde6c86aa231ff8c9b60dd667879d2890f7b33086b" exitCode=0 Nov 24 07:58:29 crc kubenswrapper[4799]: I1124 07:58:29.557268 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4","Type":"ContainerDied","Data":"5cfa07f9c8617bb6e62c98dde6c86aa231ff8c9b60dd667879d2890f7b33086b"} Nov 24 07:58:29 crc kubenswrapper[4799]: I1124 07:58:29.580231 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.580212143 podStartE2EDuration="36.580212143s" podCreationTimestamp="2025-11-24 07:57:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:58:29.573824241 +0000 UTC m=+4255.229806705" watchObservedRunningTime="2025-11-24 07:58:29.580212143 +0000 UTC m=+4255.236194617" Nov 24 07:58:30 crc kubenswrapper[4799]: I1124 07:58:30.570672 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4","Type":"ContainerStarted","Data":"931bfb86e7f2b38dd8221b7426588ea25cfb4071d186717c320c5d09cbc768b9"} Nov 24 07:58:30 crc kubenswrapper[4799]: I1124 07:58:30.571059 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:58:30 crc kubenswrapper[4799]: I1124 07:58:30.598323 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.598292218 podStartE2EDuration="37.598292218s" podCreationTimestamp="2025-11-24 07:57:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:58:30.594888071 +0000 UTC m=+4256.250870615" watchObservedRunningTime="2025-11-24 07:58:30.598292218 +0000 UTC m=+4256.254274732" Nov 24 07:58:43 crc kubenswrapper[4799]: I1124 07:58:43.582271 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 07:58:43 crc kubenswrapper[4799]: I1124 07:58:43.935063 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.400327 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.401183 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.401272 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.402378 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.402488 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" gracePeriod=600 Nov 24 07:58:50 crc kubenswrapper[4799]: E1124 07:58:50.540790 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.965661 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 07:58:50 crc kubenswrapper[4799]: E1124 07:58:50.966135 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c5f2a8-b835-4efe-9c2e-42deb39aeb66" containerName="dnsmasq-dns" Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.966164 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c5f2a8-b835-4efe-9c2e-42deb39aeb66" containerName="dnsmasq-dns" Nov 24 07:58:50 crc kubenswrapper[4799]: E1124 07:58:50.966195 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c5f2a8-b835-4efe-9c2e-42deb39aeb66" containerName="init" Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.966208 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c5f2a8-b835-4efe-9c2e-42deb39aeb66" containerName="init" Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.966499 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c5f2a8-b835-4efe-9c2e-42deb39aeb66" containerName="dnsmasq-dns" Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.967909 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.970795 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zmjxb" Nov 24 07:58:50 crc kubenswrapper[4799]: I1124 07:58:50.979844 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 07:58:51 crc kubenswrapper[4799]: I1124 07:58:51.122823 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2dsh\" (UniqueName: \"kubernetes.io/projected/faff77e9-ef4d-400a-a67a-bcf608b224d7-kube-api-access-c2dsh\") pod \"mariadb-client-1-default\" (UID: \"faff77e9-ef4d-400a-a67a-bcf608b224d7\") " pod="openstack/mariadb-client-1-default" Nov 24 07:58:51 crc kubenswrapper[4799]: I1124 07:58:51.156359 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" exitCode=0 Nov 24 07:58:51 crc kubenswrapper[4799]: I1124 07:58:51.156411 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323"} Nov 24 07:58:51 crc kubenswrapper[4799]: I1124 07:58:51.156450 4799 scope.go:117] "RemoveContainer" containerID="79a567e5468d72ae31fb5bce03a697890201831b25f9fd884a4780275b786543" Nov 24 07:58:51 crc kubenswrapper[4799]: I1124 07:58:51.157155 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 07:58:51 crc kubenswrapper[4799]: E1124 07:58:51.157383 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:58:51 crc kubenswrapper[4799]: I1124 07:58:51.224394 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2dsh\" (UniqueName: \"kubernetes.io/projected/faff77e9-ef4d-400a-a67a-bcf608b224d7-kube-api-access-c2dsh\") pod \"mariadb-client-1-default\" (UID: \"faff77e9-ef4d-400a-a67a-bcf608b224d7\") " pod="openstack/mariadb-client-1-default" Nov 24 07:58:51 crc kubenswrapper[4799]: I1124 07:58:51.262497 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2dsh\" (UniqueName: \"kubernetes.io/projected/faff77e9-ef4d-400a-a67a-bcf608b224d7-kube-api-access-c2dsh\") pod \"mariadb-client-1-default\" (UID: \"faff77e9-ef4d-400a-a67a-bcf608b224d7\") " pod="openstack/mariadb-client-1-default" Nov 24 07:58:51 crc kubenswrapper[4799]: I1124 07:58:51.288245 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 07:58:51 crc kubenswrapper[4799]: I1124 07:58:51.821228 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 07:58:52 crc kubenswrapper[4799]: I1124 07:58:52.176279 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"faff77e9-ef4d-400a-a67a-bcf608b224d7","Type":"ContainerStarted","Data":"1323d186f089ff22e1898f88427322be5a86586a9d98c221d4ab6528ba835feb"} Nov 24 07:58:53 crc kubenswrapper[4799]: I1124 07:58:53.188110 4799 generic.go:334] "Generic (PLEG): container finished" podID="faff77e9-ef4d-400a-a67a-bcf608b224d7" containerID="93629a7205f9f1bc69fa835ffa37243c0154ab48ee6e86d52d814e2dabbe445b" exitCode=0 Nov 24 07:58:53 crc kubenswrapper[4799]: I1124 07:58:53.188245 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"faff77e9-ef4d-400a-a67a-bcf608b224d7","Type":"ContainerDied","Data":"93629a7205f9f1bc69fa835ffa37243c0154ab48ee6e86d52d814e2dabbe445b"} Nov 24 07:58:54 crc kubenswrapper[4799]: I1124 07:58:54.550806 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 07:58:54 crc kubenswrapper[4799]: I1124 07:58:54.594088 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_faff77e9-ef4d-400a-a67a-bcf608b224d7/mariadb-client-1-default/0.log" Nov 24 07:58:54 crc kubenswrapper[4799]: I1124 07:58:54.618313 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 07:58:54 crc kubenswrapper[4799]: I1124 07:58:54.623577 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 07:58:54 crc kubenswrapper[4799]: I1124 07:58:54.694392 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2dsh\" (UniqueName: \"kubernetes.io/projected/faff77e9-ef4d-400a-a67a-bcf608b224d7-kube-api-access-c2dsh\") pod \"faff77e9-ef4d-400a-a67a-bcf608b224d7\" (UID: \"faff77e9-ef4d-400a-a67a-bcf608b224d7\") " Nov 24 07:58:54 crc kubenswrapper[4799]: I1124 07:58:54.700083 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faff77e9-ef4d-400a-a67a-bcf608b224d7-kube-api-access-c2dsh" (OuterVolumeSpecName: "kube-api-access-c2dsh") pod "faff77e9-ef4d-400a-a67a-bcf608b224d7" (UID: "faff77e9-ef4d-400a-a67a-bcf608b224d7"). InnerVolumeSpecName "kube-api-access-c2dsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:58:54 crc kubenswrapper[4799]: I1124 07:58:54.796555 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2dsh\" (UniqueName: \"kubernetes.io/projected/faff77e9-ef4d-400a-a67a-bcf608b224d7-kube-api-access-c2dsh\") on node \"crc\" DevicePath \"\"" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.156752 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 07:58:55 crc kubenswrapper[4799]: E1124 07:58:55.157613 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faff77e9-ef4d-400a-a67a-bcf608b224d7" containerName="mariadb-client-1-default" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.157656 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="faff77e9-ef4d-400a-a67a-bcf608b224d7" containerName="mariadb-client-1-default" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.158079 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="faff77e9-ef4d-400a-a67a-bcf608b224d7" containerName="mariadb-client-1-default" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.158979 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.168946 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.213652 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1323d186f089ff22e1898f88427322be5a86586a9d98c221d4ab6528ba835feb" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.213755 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.304719 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7chr\" (UniqueName: \"kubernetes.io/projected/a70701cb-8c8a-4467-bd1a-5152ea5b13e8-kube-api-access-x7chr\") pod \"mariadb-client-2-default\" (UID: \"a70701cb-8c8a-4467-bd1a-5152ea5b13e8\") " pod="openstack/mariadb-client-2-default" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.406149 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7chr\" (UniqueName: \"kubernetes.io/projected/a70701cb-8c8a-4467-bd1a-5152ea5b13e8-kube-api-access-x7chr\") pod \"mariadb-client-2-default\" (UID: \"a70701cb-8c8a-4467-bd1a-5152ea5b13e8\") " pod="openstack/mariadb-client-2-default" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.425956 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7chr\" (UniqueName: \"kubernetes.io/projected/a70701cb-8c8a-4467-bd1a-5152ea5b13e8-kube-api-access-x7chr\") pod \"mariadb-client-2-default\" (UID: \"a70701cb-8c8a-4467-bd1a-5152ea5b13e8\") " pod="openstack/mariadb-client-2-default" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.492825 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.655463 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faff77e9-ef4d-400a-a67a-bcf608b224d7" path="/var/lib/kubelet/pods/faff77e9-ef4d-400a-a67a-bcf608b224d7/volumes" Nov 24 07:58:55 crc kubenswrapper[4799]: I1124 07:58:55.938165 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 07:58:56 crc kubenswrapper[4799]: I1124 07:58:56.223807 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"a70701cb-8c8a-4467-bd1a-5152ea5b13e8","Type":"ContainerStarted","Data":"0356867f9d47ad29a15da72d5d230d72c1ded27aed899d8a9e0a67454eea719f"} Nov 24 07:58:56 crc kubenswrapper[4799]: I1124 07:58:56.224179 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"a70701cb-8c8a-4467-bd1a-5152ea5b13e8","Type":"ContainerStarted","Data":"f40587e0ef3b8228a0c56f3c311b69e4491c51b5c311b36f0331501131918b5b"} Nov 24 07:58:56 crc kubenswrapper[4799]: I1124 07:58:56.243085 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=1.243065351 podStartE2EDuration="1.243065351s" podCreationTimestamp="2025-11-24 07:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:58:56.23952274 +0000 UTC m=+4281.895505214" watchObservedRunningTime="2025-11-24 07:58:56.243065351 +0000 UTC m=+4281.899047825" Nov 24 07:58:57 crc kubenswrapper[4799]: I1124 07:58:57.234234 4799 generic.go:334] "Generic (PLEG): container finished" podID="a70701cb-8c8a-4467-bd1a-5152ea5b13e8" containerID="0356867f9d47ad29a15da72d5d230d72c1ded27aed899d8a9e0a67454eea719f" exitCode=1 Nov 24 07:58:57 crc kubenswrapper[4799]: I1124 07:58:57.234308 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"a70701cb-8c8a-4467-bd1a-5152ea5b13e8","Type":"ContainerDied","Data":"0356867f9d47ad29a15da72d5d230d72c1ded27aed899d8a9e0a67454eea719f"} Nov 24 07:58:58 crc kubenswrapper[4799]: I1124 07:58:58.679375 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 07:58:58 crc kubenswrapper[4799]: I1124 07:58:58.720893 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 07:58:58 crc kubenswrapper[4799]: I1124 07:58:58.728476 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 07:58:58 crc kubenswrapper[4799]: I1124 07:58:58.868277 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7chr\" (UniqueName: \"kubernetes.io/projected/a70701cb-8c8a-4467-bd1a-5152ea5b13e8-kube-api-access-x7chr\") pod \"a70701cb-8c8a-4467-bd1a-5152ea5b13e8\" (UID: \"a70701cb-8c8a-4467-bd1a-5152ea5b13e8\") " Nov 24 07:58:58 crc kubenswrapper[4799]: I1124 07:58:58.884655 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a70701cb-8c8a-4467-bd1a-5152ea5b13e8-kube-api-access-x7chr" (OuterVolumeSpecName: "kube-api-access-x7chr") pod "a70701cb-8c8a-4467-bd1a-5152ea5b13e8" (UID: "a70701cb-8c8a-4467-bd1a-5152ea5b13e8"). InnerVolumeSpecName "kube-api-access-x7chr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:58:58 crc kubenswrapper[4799]: I1124 07:58:58.970514 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7chr\" (UniqueName: \"kubernetes.io/projected/a70701cb-8c8a-4467-bd1a-5152ea5b13e8-kube-api-access-x7chr\") on node \"crc\" DevicePath \"\"" Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.242062 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Nov 24 07:58:59 crc kubenswrapper[4799]: E1124 07:58:59.243129 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70701cb-8c8a-4467-bd1a-5152ea5b13e8" containerName="mariadb-client-2-default" Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.243175 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70701cb-8c8a-4467-bd1a-5152ea5b13e8" containerName="mariadb-client-2-default" Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.243541 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a70701cb-8c8a-4467-bd1a-5152ea5b13e8" containerName="mariadb-client-2-default" Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.244778 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.252460 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.281093 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f40587e0ef3b8228a0c56f3c311b69e4491c51b5c311b36f0331501131918b5b" Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.281338 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.379359 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz84g\" (UniqueName: \"kubernetes.io/projected/dcd1f9ac-d567-483c-aa41-295096fb51e1-kube-api-access-rz84g\") pod \"mariadb-client-1\" (UID: \"dcd1f9ac-d567-483c-aa41-295096fb51e1\") " pod="openstack/mariadb-client-1" Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.481495 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz84g\" (UniqueName: \"kubernetes.io/projected/dcd1f9ac-d567-483c-aa41-295096fb51e1-kube-api-access-rz84g\") pod \"mariadb-client-1\" (UID: \"dcd1f9ac-d567-483c-aa41-295096fb51e1\") " pod="openstack/mariadb-client-1" Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.512064 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz84g\" (UniqueName: \"kubernetes.io/projected/dcd1f9ac-d567-483c-aa41-295096fb51e1-kube-api-access-rz84g\") pod \"mariadb-client-1\" (UID: \"dcd1f9ac-d567-483c-aa41-295096fb51e1\") " pod="openstack/mariadb-client-1" Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.580196 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 07:58:59 crc kubenswrapper[4799]: I1124 07:58:59.653362 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a70701cb-8c8a-4467-bd1a-5152ea5b13e8" path="/var/lib/kubelet/pods/a70701cb-8c8a-4467-bd1a-5152ea5b13e8/volumes" Nov 24 07:59:00 crc kubenswrapper[4799]: I1124 07:59:00.162764 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 07:59:00 crc kubenswrapper[4799]: W1124 07:59:00.169710 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddcd1f9ac_d567_483c_aa41_295096fb51e1.slice/crio-12ac7c07cffc968713b9545d68925ed613913bd13c8e3c65b24178136fa17aee WatchSource:0}: Error finding container 12ac7c07cffc968713b9545d68925ed613913bd13c8e3c65b24178136fa17aee: Status 404 returned error can't find the container with id 12ac7c07cffc968713b9545d68925ed613913bd13c8e3c65b24178136fa17aee Nov 24 07:59:00 crc kubenswrapper[4799]: I1124 07:59:00.291505 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"dcd1f9ac-d567-483c-aa41-295096fb51e1","Type":"ContainerStarted","Data":"12ac7c07cffc968713b9545d68925ed613913bd13c8e3c65b24178136fa17aee"} Nov 24 07:59:01 crc kubenswrapper[4799]: I1124 07:59:01.304247 4799 generic.go:334] "Generic (PLEG): container finished" podID="dcd1f9ac-d567-483c-aa41-295096fb51e1" containerID="6de72d6906d6946f26d69d189df9b8cad79db64be0f5da3ad22cc3770925d5c9" exitCode=0 Nov 24 07:59:01 crc kubenswrapper[4799]: I1124 07:59:01.304368 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"dcd1f9ac-d567-483c-aa41-295096fb51e1","Type":"ContainerDied","Data":"6de72d6906d6946f26d69d189df9b8cad79db64be0f5da3ad22cc3770925d5c9"} Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.065473 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.091824 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_dcd1f9ac-d567-483c-aa41-295096fb51e1/mariadb-client-1/0.log" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.126924 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.135681 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.244665 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz84g\" (UniqueName: \"kubernetes.io/projected/dcd1f9ac-d567-483c-aa41-295096fb51e1-kube-api-access-rz84g\") pod \"dcd1f9ac-d567-483c-aa41-295096fb51e1\" (UID: \"dcd1f9ac-d567-483c-aa41-295096fb51e1\") " Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.251163 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcd1f9ac-d567-483c-aa41-295096fb51e1-kube-api-access-rz84g" (OuterVolumeSpecName: "kube-api-access-rz84g") pod "dcd1f9ac-d567-483c-aa41-295096fb51e1" (UID: "dcd1f9ac-d567-483c-aa41-295096fb51e1"). InnerVolumeSpecName "kube-api-access-rz84g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.327190 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12ac7c07cffc968713b9545d68925ed613913bd13c8e3c65b24178136fa17aee" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.327245 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.347157 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz84g\" (UniqueName: \"kubernetes.io/projected/dcd1f9ac-d567-483c-aa41-295096fb51e1-kube-api-access-rz84g\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.576475 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 07:59:03 crc kubenswrapper[4799]: E1124 07:59:03.576942 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcd1f9ac-d567-483c-aa41-295096fb51e1" containerName="mariadb-client-1" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.576972 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcd1f9ac-d567-483c-aa41-295096fb51e1" containerName="mariadb-client-1" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.577280 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcd1f9ac-d567-483c-aa41-295096fb51e1" containerName="mariadb-client-1" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.579907 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.584463 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zmjxb" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.585686 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.628351 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 07:59:03 crc kubenswrapper[4799]: E1124 07:59:03.628524 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.641351 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcd1f9ac-d567-483c-aa41-295096fb51e1" path="/var/lib/kubelet/pods/dcd1f9ac-d567-483c-aa41-295096fb51e1/volumes" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.753995 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qpb2\" (UniqueName: \"kubernetes.io/projected/ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea-kube-api-access-5qpb2\") pod \"mariadb-client-4-default\" (UID: \"ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea\") " pod="openstack/mariadb-client-4-default" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.855331 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qpb2\" (UniqueName: \"kubernetes.io/projected/ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea-kube-api-access-5qpb2\") pod \"mariadb-client-4-default\" (UID: \"ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea\") " pod="openstack/mariadb-client-4-default" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.874638 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qpb2\" (UniqueName: \"kubernetes.io/projected/ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea-kube-api-access-5qpb2\") pod \"mariadb-client-4-default\" (UID: \"ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea\") " pod="openstack/mariadb-client-4-default" Nov 24 07:59:03 crc kubenswrapper[4799]: I1124 07:59:03.916262 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 07:59:04 crc kubenswrapper[4799]: I1124 07:59:04.519619 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 07:59:05 crc kubenswrapper[4799]: I1124 07:59:05.346534 4799 generic.go:334] "Generic (PLEG): container finished" podID="ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea" containerID="8ed399e008d8244f6f977ed7d84630cfb86a2e4d4d3dfde3f3cf0049e52f6ad4" exitCode=0 Nov 24 07:59:05 crc kubenswrapper[4799]: I1124 07:59:05.346766 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea","Type":"ContainerDied","Data":"8ed399e008d8244f6f977ed7d84630cfb86a2e4d4d3dfde3f3cf0049e52f6ad4"} Nov 24 07:59:05 crc kubenswrapper[4799]: I1124 07:59:05.347264 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea","Type":"ContainerStarted","Data":"e50b88391c3536fdf1f7159c590dbcedcf304e8feb95d1d4af882513ccf9cea9"} Nov 24 07:59:06 crc kubenswrapper[4799]: I1124 07:59:06.791552 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 07:59:06 crc kubenswrapper[4799]: I1124 07:59:06.814887 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea/mariadb-client-4-default/0.log" Nov 24 07:59:06 crc kubenswrapper[4799]: I1124 07:59:06.841776 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 07:59:06 crc kubenswrapper[4799]: I1124 07:59:06.855926 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 07:59:06 crc kubenswrapper[4799]: I1124 07:59:06.911895 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qpb2\" (UniqueName: \"kubernetes.io/projected/ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea-kube-api-access-5qpb2\") pod \"ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea\" (UID: \"ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea\") " Nov 24 07:59:06 crc kubenswrapper[4799]: I1124 07:59:06.918571 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea-kube-api-access-5qpb2" (OuterVolumeSpecName: "kube-api-access-5qpb2") pod "ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea" (UID: "ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea"). InnerVolumeSpecName "kube-api-access-5qpb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:59:07 crc kubenswrapper[4799]: I1124 07:59:07.014089 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qpb2\" (UniqueName: \"kubernetes.io/projected/ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea-kube-api-access-5qpb2\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:07 crc kubenswrapper[4799]: I1124 07:59:07.369918 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e50b88391c3536fdf1f7159c590dbcedcf304e8feb95d1d4af882513ccf9cea9" Nov 24 07:59:07 crc kubenswrapper[4799]: I1124 07:59:07.370108 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 07:59:07 crc kubenswrapper[4799]: I1124 07:59:07.644958 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea" path="/var/lib/kubelet/pods/ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea/volumes" Nov 24 07:59:10 crc kubenswrapper[4799]: I1124 07:59:10.159125 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 07:59:10 crc kubenswrapper[4799]: E1124 07:59:10.160147 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea" containerName="mariadb-client-4-default" Nov 24 07:59:10 crc kubenswrapper[4799]: I1124 07:59:10.160182 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea" containerName="mariadb-client-4-default" Nov 24 07:59:10 crc kubenswrapper[4799]: I1124 07:59:10.160504 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce7d3bc5-cdb4-4fcc-bb73-3fbbae8e0aea" containerName="mariadb-client-4-default" Nov 24 07:59:10 crc kubenswrapper[4799]: I1124 07:59:10.162988 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 07:59:10 crc kubenswrapper[4799]: I1124 07:59:10.165132 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 07:59:10 crc kubenswrapper[4799]: I1124 07:59:10.165757 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zmjxb" Nov 24 07:59:10 crc kubenswrapper[4799]: I1124 07:59:10.284283 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvdvk\" (UniqueName: \"kubernetes.io/projected/beb32413-46a7-4fdf-97b8-0ab125a2bbd1-kube-api-access-cvdvk\") pod \"mariadb-client-5-default\" (UID: \"beb32413-46a7-4fdf-97b8-0ab125a2bbd1\") " pod="openstack/mariadb-client-5-default" Nov 24 07:59:10 crc kubenswrapper[4799]: I1124 07:59:10.385658 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvdvk\" (UniqueName: \"kubernetes.io/projected/beb32413-46a7-4fdf-97b8-0ab125a2bbd1-kube-api-access-cvdvk\") pod \"mariadb-client-5-default\" (UID: \"beb32413-46a7-4fdf-97b8-0ab125a2bbd1\") " pod="openstack/mariadb-client-5-default" Nov 24 07:59:10 crc kubenswrapper[4799]: I1124 07:59:10.418806 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvdvk\" (UniqueName: \"kubernetes.io/projected/beb32413-46a7-4fdf-97b8-0ab125a2bbd1-kube-api-access-cvdvk\") pod \"mariadb-client-5-default\" (UID: \"beb32413-46a7-4fdf-97b8-0ab125a2bbd1\") " pod="openstack/mariadb-client-5-default" Nov 24 07:59:10 crc kubenswrapper[4799]: I1124 07:59:10.496601 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 07:59:10 crc kubenswrapper[4799]: I1124 07:59:10.828746 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 07:59:11 crc kubenswrapper[4799]: I1124 07:59:11.406559 4799 generic.go:334] "Generic (PLEG): container finished" podID="beb32413-46a7-4fdf-97b8-0ab125a2bbd1" containerID="e8573b08f19e25d030ab4588a0906815dc5827afa37b711ab2339e8e2da65a9f" exitCode=0 Nov 24 07:59:11 crc kubenswrapper[4799]: I1124 07:59:11.406618 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"beb32413-46a7-4fdf-97b8-0ab125a2bbd1","Type":"ContainerDied","Data":"e8573b08f19e25d030ab4588a0906815dc5827afa37b711ab2339e8e2da65a9f"} Nov 24 07:59:11 crc kubenswrapper[4799]: I1124 07:59:11.406935 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"beb32413-46a7-4fdf-97b8-0ab125a2bbd1","Type":"ContainerStarted","Data":"484b2a11185d4200a82f68043f3c0e7f82bca74b1d5d20b1dd3c30aca363ec66"} Nov 24 07:59:12 crc kubenswrapper[4799]: I1124 07:59:12.818125 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 07:59:12 crc kubenswrapper[4799]: I1124 07:59:12.823513 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvdvk\" (UniqueName: \"kubernetes.io/projected/beb32413-46a7-4fdf-97b8-0ab125a2bbd1-kube-api-access-cvdvk\") pod \"beb32413-46a7-4fdf-97b8-0ab125a2bbd1\" (UID: \"beb32413-46a7-4fdf-97b8-0ab125a2bbd1\") " Nov 24 07:59:12 crc kubenswrapper[4799]: I1124 07:59:12.829174 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beb32413-46a7-4fdf-97b8-0ab125a2bbd1-kube-api-access-cvdvk" (OuterVolumeSpecName: "kube-api-access-cvdvk") pod "beb32413-46a7-4fdf-97b8-0ab125a2bbd1" (UID: "beb32413-46a7-4fdf-97b8-0ab125a2bbd1"). InnerVolumeSpecName "kube-api-access-cvdvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:59:12 crc kubenswrapper[4799]: I1124 07:59:12.860029 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_beb32413-46a7-4fdf-97b8-0ab125a2bbd1/mariadb-client-5-default/0.log" Nov 24 07:59:12 crc kubenswrapper[4799]: I1124 07:59:12.904017 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 07:59:12 crc kubenswrapper[4799]: I1124 07:59:12.909192 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 07:59:12 crc kubenswrapper[4799]: I1124 07:59:12.924943 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvdvk\" (UniqueName: \"kubernetes.io/projected/beb32413-46a7-4fdf-97b8-0ab125a2bbd1-kube-api-access-cvdvk\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.035832 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 07:59:13 crc kubenswrapper[4799]: E1124 07:59:13.036284 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb32413-46a7-4fdf-97b8-0ab125a2bbd1" containerName="mariadb-client-5-default" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.036306 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb32413-46a7-4fdf-97b8-0ab125a2bbd1" containerName="mariadb-client-5-default" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.036521 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb32413-46a7-4fdf-97b8-0ab125a2bbd1" containerName="mariadb-client-5-default" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.037362 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.041298 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.129995 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj76k\" (UniqueName: \"kubernetes.io/projected/06f72cdb-b804-4a5e-bf84-f6fafc860b20-kube-api-access-xj76k\") pod \"mariadb-client-6-default\" (UID: \"06f72cdb-b804-4a5e-bf84-f6fafc860b20\") " pod="openstack/mariadb-client-6-default" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.231245 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj76k\" (UniqueName: \"kubernetes.io/projected/06f72cdb-b804-4a5e-bf84-f6fafc860b20-kube-api-access-xj76k\") pod \"mariadb-client-6-default\" (UID: \"06f72cdb-b804-4a5e-bf84-f6fafc860b20\") " pod="openstack/mariadb-client-6-default" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.262920 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj76k\" (UniqueName: \"kubernetes.io/projected/06f72cdb-b804-4a5e-bf84-f6fafc860b20-kube-api-access-xj76k\") pod \"mariadb-client-6-default\" (UID: \"06f72cdb-b804-4a5e-bf84-f6fafc860b20\") " pod="openstack/mariadb-client-6-default" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.365879 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.434723 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="484b2a11185d4200a82f68043f3c0e7f82bca74b1d5d20b1dd3c30aca363ec66" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.434822 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.645565 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beb32413-46a7-4fdf-97b8-0ab125a2bbd1" path="/var/lib/kubelet/pods/beb32413-46a7-4fdf-97b8-0ab125a2bbd1/volumes" Nov 24 07:59:13 crc kubenswrapper[4799]: I1124 07:59:13.784951 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 07:59:13 crc kubenswrapper[4799]: W1124 07:59:13.793460 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06f72cdb_b804_4a5e_bf84_f6fafc860b20.slice/crio-c9fb15973b2767eb779e8792c8f383f04c71fd45c1bbd65f64657681e2f8b753 WatchSource:0}: Error finding container c9fb15973b2767eb779e8792c8f383f04c71fd45c1bbd65f64657681e2f8b753: Status 404 returned error can't find the container with id c9fb15973b2767eb779e8792c8f383f04c71fd45c1bbd65f64657681e2f8b753 Nov 24 07:59:14 crc kubenswrapper[4799]: I1124 07:59:14.447198 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"06f72cdb-b804-4a5e-bf84-f6fafc860b20","Type":"ContainerStarted","Data":"0c7e10862a39a5a52d591852b7ce843d5388fb7dbccf635e685b637b483b1a59"} Nov 24 07:59:14 crc kubenswrapper[4799]: I1124 07:59:14.447686 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"06f72cdb-b804-4a5e-bf84-f6fafc860b20","Type":"ContainerStarted","Data":"c9fb15973b2767eb779e8792c8f383f04c71fd45c1bbd65f64657681e2f8b753"} Nov 24 07:59:14 crc kubenswrapper[4799]: I1124 07:59:14.477560 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=1.477530048 podStartE2EDuration="1.477530048s" podCreationTimestamp="2025-11-24 07:59:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 07:59:14.467834493 +0000 UTC m=+4300.123816997" watchObservedRunningTime="2025-11-24 07:59:14.477530048 +0000 UTC m=+4300.133512522" Nov 24 07:59:14 crc kubenswrapper[4799]: I1124 07:59:14.521766 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_06f72cdb-b804-4a5e-bf84-f6fafc860b20/mariadb-client-6-default/0.log" Nov 24 07:59:15 crc kubenswrapper[4799]: I1124 07:59:15.458536 4799 generic.go:334] "Generic (PLEG): container finished" podID="06f72cdb-b804-4a5e-bf84-f6fafc860b20" containerID="0c7e10862a39a5a52d591852b7ce843d5388fb7dbccf635e685b637b483b1a59" exitCode=1 Nov 24 07:59:15 crc kubenswrapper[4799]: I1124 07:59:15.458678 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"06f72cdb-b804-4a5e-bf84-f6fafc860b20","Type":"ContainerDied","Data":"0c7e10862a39a5a52d591852b7ce843d5388fb7dbccf635e685b637b483b1a59"} Nov 24 07:59:16 crc kubenswrapper[4799]: I1124 07:59:16.896298 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 07:59:16 crc kubenswrapper[4799]: I1124 07:59:16.965508 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 07:59:16 crc kubenswrapper[4799]: I1124 07:59:16.979618 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.001689 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xj76k\" (UniqueName: \"kubernetes.io/projected/06f72cdb-b804-4a5e-bf84-f6fafc860b20-kube-api-access-xj76k\") pod \"06f72cdb-b804-4a5e-bf84-f6fafc860b20\" (UID: \"06f72cdb-b804-4a5e-bf84-f6fafc860b20\") " Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.006564 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f72cdb-b804-4a5e-bf84-f6fafc860b20-kube-api-access-xj76k" (OuterVolumeSpecName: "kube-api-access-xj76k") pod "06f72cdb-b804-4a5e-bf84-f6fafc860b20" (UID: "06f72cdb-b804-4a5e-bf84-f6fafc860b20"). InnerVolumeSpecName "kube-api-access-xj76k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.104294 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xj76k\" (UniqueName: \"kubernetes.io/projected/06f72cdb-b804-4a5e-bf84-f6fafc860b20-kube-api-access-xj76k\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.106944 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 07:59:17 crc kubenswrapper[4799]: E1124 07:59:17.107331 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f72cdb-b804-4a5e-bf84-f6fafc860b20" containerName="mariadb-client-6-default" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.107348 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f72cdb-b804-4a5e-bf84-f6fafc860b20" containerName="mariadb-client-6-default" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.107547 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f72cdb-b804-4a5e-bf84-f6fafc860b20" containerName="mariadb-client-6-default" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.108186 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.123209 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.306821 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fd7z\" (UniqueName: \"kubernetes.io/projected/7ef7a56c-b173-4ff9-8377-145de40ed2a3-kube-api-access-6fd7z\") pod \"mariadb-client-7-default\" (UID: \"7ef7a56c-b173-4ff9-8377-145de40ed2a3\") " pod="openstack/mariadb-client-7-default" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.408083 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fd7z\" (UniqueName: \"kubernetes.io/projected/7ef7a56c-b173-4ff9-8377-145de40ed2a3-kube-api-access-6fd7z\") pod \"mariadb-client-7-default\" (UID: \"7ef7a56c-b173-4ff9-8377-145de40ed2a3\") " pod="openstack/mariadb-client-7-default" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.438279 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fd7z\" (UniqueName: \"kubernetes.io/projected/7ef7a56c-b173-4ff9-8377-145de40ed2a3-kube-api-access-6fd7z\") pod \"mariadb-client-7-default\" (UID: \"7ef7a56c-b173-4ff9-8377-145de40ed2a3\") " pod="openstack/mariadb-client-7-default" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.481895 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9fb15973b2767eb779e8792c8f383f04c71fd45c1bbd65f64657681e2f8b753" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.482120 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.646018 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f72cdb-b804-4a5e-bf84-f6fafc860b20" path="/var/lib/kubelet/pods/06f72cdb-b804-4a5e-bf84-f6fafc860b20/volumes" Nov 24 07:59:17 crc kubenswrapper[4799]: I1124 07:59:17.724733 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 07:59:18 crc kubenswrapper[4799]: I1124 07:59:18.334689 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 07:59:18 crc kubenswrapper[4799]: I1124 07:59:18.491805 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"7ef7a56c-b173-4ff9-8377-145de40ed2a3","Type":"ContainerStarted","Data":"5c1501ac51f650c0d3acf34264ecc9f72c022749815eda017a6750a3a35c2e62"} Nov 24 07:59:18 crc kubenswrapper[4799]: I1124 07:59:18.628050 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 07:59:18 crc kubenswrapper[4799]: E1124 07:59:18.628301 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:59:19 crc kubenswrapper[4799]: I1124 07:59:19.506389 4799 generic.go:334] "Generic (PLEG): container finished" podID="7ef7a56c-b173-4ff9-8377-145de40ed2a3" containerID="d33c246e9b2dc9a3e6c2b9c329f572a2598bbc9753f54f639bcefc377a192cb4" exitCode=0 Nov 24 07:59:19 crc kubenswrapper[4799]: I1124 07:59:19.506689 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"7ef7a56c-b173-4ff9-8377-145de40ed2a3","Type":"ContainerDied","Data":"d33c246e9b2dc9a3e6c2b9c329f572a2598bbc9753f54f639bcefc377a192cb4"} Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.071923 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.095282 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_7ef7a56c-b173-4ff9-8377-145de40ed2a3/mariadb-client-7-default/0.log" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.123767 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.128405 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.178662 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fd7z\" (UniqueName: \"kubernetes.io/projected/7ef7a56c-b173-4ff9-8377-145de40ed2a3-kube-api-access-6fd7z\") pod \"7ef7a56c-b173-4ff9-8377-145de40ed2a3\" (UID: \"7ef7a56c-b173-4ff9-8377-145de40ed2a3\") " Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.183746 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ef7a56c-b173-4ff9-8377-145de40ed2a3-kube-api-access-6fd7z" (OuterVolumeSpecName: "kube-api-access-6fd7z") pod "7ef7a56c-b173-4ff9-8377-145de40ed2a3" (UID: "7ef7a56c-b173-4ff9-8377-145de40ed2a3"). InnerVolumeSpecName "kube-api-access-6fd7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.250309 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Nov 24 07:59:21 crc kubenswrapper[4799]: E1124 07:59:21.250599 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ef7a56c-b173-4ff9-8377-145de40ed2a3" containerName="mariadb-client-7-default" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.250614 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ef7a56c-b173-4ff9-8377-145de40ed2a3" containerName="mariadb-client-7-default" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.250796 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ef7a56c-b173-4ff9-8377-145de40ed2a3" containerName="mariadb-client-7-default" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.251309 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.262489 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.280404 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fd7z\" (UniqueName: \"kubernetes.io/projected/7ef7a56c-b173-4ff9-8377-145de40ed2a3-kube-api-access-6fd7z\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.381537 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r8h5\" (UniqueName: \"kubernetes.io/projected/9e88406e-4fc1-4fe8-b5bb-04bd49030672-kube-api-access-6r8h5\") pod \"mariadb-client-2\" (UID: \"9e88406e-4fc1-4fe8-b5bb-04bd49030672\") " pod="openstack/mariadb-client-2" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.483081 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r8h5\" (UniqueName: \"kubernetes.io/projected/9e88406e-4fc1-4fe8-b5bb-04bd49030672-kube-api-access-6r8h5\") pod \"mariadb-client-2\" (UID: \"9e88406e-4fc1-4fe8-b5bb-04bd49030672\") " pod="openstack/mariadb-client-2" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.513428 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r8h5\" (UniqueName: \"kubernetes.io/projected/9e88406e-4fc1-4fe8-b5bb-04bd49030672-kube-api-access-6r8h5\") pod \"mariadb-client-2\" (UID: \"9e88406e-4fc1-4fe8-b5bb-04bd49030672\") " pod="openstack/mariadb-client-2" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.526090 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c1501ac51f650c0d3acf34264ecc9f72c022749815eda017a6750a3a35c2e62" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.526156 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.570228 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.640163 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ef7a56c-b173-4ff9-8377-145de40ed2a3" path="/var/lib/kubelet/pods/7ef7a56c-b173-4ff9-8377-145de40ed2a3/volumes" Nov 24 07:59:21 crc kubenswrapper[4799]: I1124 07:59:21.906308 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 07:59:21 crc kubenswrapper[4799]: W1124 07:59:21.918527 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e88406e_4fc1_4fe8_b5bb_04bd49030672.slice/crio-fc308bcea8a06434b4551553c1da8ef820d2352ff6bf8832e29e076a94ed4058 WatchSource:0}: Error finding container fc308bcea8a06434b4551553c1da8ef820d2352ff6bf8832e29e076a94ed4058: Status 404 returned error can't find the container with id fc308bcea8a06434b4551553c1da8ef820d2352ff6bf8832e29e076a94ed4058 Nov 24 07:59:22 crc kubenswrapper[4799]: I1124 07:59:22.540563 4799 generic.go:334] "Generic (PLEG): container finished" podID="9e88406e-4fc1-4fe8-b5bb-04bd49030672" containerID="36326a5305ebb27050b9faeb368319a3091f2989685d5a9009fb90dd71518b8b" exitCode=0 Nov 24 07:59:22 crc kubenswrapper[4799]: I1124 07:59:22.540746 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"9e88406e-4fc1-4fe8-b5bb-04bd49030672","Type":"ContainerDied","Data":"36326a5305ebb27050b9faeb368319a3091f2989685d5a9009fb90dd71518b8b"} Nov 24 07:59:22 crc kubenswrapper[4799]: I1124 07:59:22.541041 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"9e88406e-4fc1-4fe8-b5bb-04bd49030672","Type":"ContainerStarted","Data":"fc308bcea8a06434b4551553c1da8ef820d2352ff6bf8832e29e076a94ed4058"} Nov 24 07:59:24 crc kubenswrapper[4799]: I1124 07:59:24.034662 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 07:59:24 crc kubenswrapper[4799]: I1124 07:59:24.052202 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_9e88406e-4fc1-4fe8-b5bb-04bd49030672/mariadb-client-2/0.log" Nov 24 07:59:24 crc kubenswrapper[4799]: I1124 07:59:24.076259 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 07:59:24 crc kubenswrapper[4799]: I1124 07:59:24.083816 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 07:59:24 crc kubenswrapper[4799]: I1124 07:59:24.223995 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r8h5\" (UniqueName: \"kubernetes.io/projected/9e88406e-4fc1-4fe8-b5bb-04bd49030672-kube-api-access-6r8h5\") pod \"9e88406e-4fc1-4fe8-b5bb-04bd49030672\" (UID: \"9e88406e-4fc1-4fe8-b5bb-04bd49030672\") " Nov 24 07:59:24 crc kubenswrapper[4799]: I1124 07:59:24.232770 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e88406e-4fc1-4fe8-b5bb-04bd49030672-kube-api-access-6r8h5" (OuterVolumeSpecName: "kube-api-access-6r8h5") pod "9e88406e-4fc1-4fe8-b5bb-04bd49030672" (UID: "9e88406e-4fc1-4fe8-b5bb-04bd49030672"). InnerVolumeSpecName "kube-api-access-6r8h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 07:59:24 crc kubenswrapper[4799]: I1124 07:59:24.326186 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r8h5\" (UniqueName: \"kubernetes.io/projected/9e88406e-4fc1-4fe8-b5bb-04bd49030672-kube-api-access-6r8h5\") on node \"crc\" DevicePath \"\"" Nov 24 07:59:24 crc kubenswrapper[4799]: I1124 07:59:24.569336 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc308bcea8a06434b4551553c1da8ef820d2352ff6bf8832e29e076a94ed4058" Nov 24 07:59:24 crc kubenswrapper[4799]: I1124 07:59:24.569843 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 07:59:25 crc kubenswrapper[4799]: I1124 07:59:25.646426 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e88406e-4fc1-4fe8-b5bb-04bd49030672" path="/var/lib/kubelet/pods/9e88406e-4fc1-4fe8-b5bb-04bd49030672/volumes" Nov 24 07:59:29 crc kubenswrapper[4799]: I1124 07:59:29.628270 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 07:59:29 crc kubenswrapper[4799]: E1124 07:59:29.628512 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:59:43 crc kubenswrapper[4799]: I1124 07:59:43.627792 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 07:59:43 crc kubenswrapper[4799]: E1124 07:59:43.628642 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 07:59:54 crc kubenswrapper[4799]: I1124 07:59:54.226623 4799 scope.go:117] "RemoveContainer" containerID="2b06ede5a007b86e84f723b97557a301b6f40e96cb94c638f67ade122b9bfd2a" Nov 24 07:59:56 crc kubenswrapper[4799]: I1124 07:59:56.629413 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 07:59:56 crc kubenswrapper[4799]: E1124 07:59:56.630194 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.155586 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc"] Nov 24 08:00:00 crc kubenswrapper[4799]: E1124 08:00:00.156319 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e88406e-4fc1-4fe8-b5bb-04bd49030672" containerName="mariadb-client-2" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.156336 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e88406e-4fc1-4fe8-b5bb-04bd49030672" containerName="mariadb-client-2" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.156587 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e88406e-4fc1-4fe8-b5bb-04bd49030672" containerName="mariadb-client-2" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.157394 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.159720 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.160444 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.168195 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc"] Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.191104 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-config-volume\") pod \"collect-profiles-29399520-p4wbc\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.191193 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-secret-volume\") pod \"collect-profiles-29399520-p4wbc\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.191237 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb2pd\" (UniqueName: \"kubernetes.io/projected/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-kube-api-access-fb2pd\") pod \"collect-profiles-29399520-p4wbc\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.293046 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-config-volume\") pod \"collect-profiles-29399520-p4wbc\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.293123 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-secret-volume\") pod \"collect-profiles-29399520-p4wbc\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.293170 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb2pd\" (UniqueName: \"kubernetes.io/projected/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-kube-api-access-fb2pd\") pod \"collect-profiles-29399520-p4wbc\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.294283 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-config-volume\") pod \"collect-profiles-29399520-p4wbc\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.299607 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-secret-volume\") pod \"collect-profiles-29399520-p4wbc\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.310396 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb2pd\" (UniqueName: \"kubernetes.io/projected/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-kube-api-access-fb2pd\") pod \"collect-profiles-29399520-p4wbc\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.485372 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.730222 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc"] Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.936330 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" event={"ID":"9e1170b5-6a86-479d-9f29-dcffda7f5ba7","Type":"ContainerStarted","Data":"8b11a0a33522def7c70ba09d3f52e193848422479019be5c27d2278e307c7b3d"} Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.936988 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" event={"ID":"9e1170b5-6a86-479d-9f29-dcffda7f5ba7","Type":"ContainerStarted","Data":"438f26f1c5c44d4244993201bb7a205d9979c7d36bfec1f450cafaf7860a6419"} Nov 24 08:00:00 crc kubenswrapper[4799]: I1124 08:00:00.958429 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" podStartSLOduration=0.958409367 podStartE2EDuration="958.409367ms" podCreationTimestamp="2025-11-24 08:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:00:00.9543051 +0000 UTC m=+4346.610287604" watchObservedRunningTime="2025-11-24 08:00:00.958409367 +0000 UTC m=+4346.614391841" Nov 24 08:00:01 crc kubenswrapper[4799]: I1124 08:00:01.946636 4799 generic.go:334] "Generic (PLEG): container finished" podID="9e1170b5-6a86-479d-9f29-dcffda7f5ba7" containerID="8b11a0a33522def7c70ba09d3f52e193848422479019be5c27d2278e307c7b3d" exitCode=0 Nov 24 08:00:01 crc kubenswrapper[4799]: I1124 08:00:01.946682 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" event={"ID":"9e1170b5-6a86-479d-9f29-dcffda7f5ba7","Type":"ContainerDied","Data":"8b11a0a33522def7c70ba09d3f52e193848422479019be5c27d2278e307c7b3d"} Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.332073 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.439799 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb2pd\" (UniqueName: \"kubernetes.io/projected/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-kube-api-access-fb2pd\") pod \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.439927 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-config-volume\") pod \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.440090 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-secret-volume\") pod \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\" (UID: \"9e1170b5-6a86-479d-9f29-dcffda7f5ba7\") " Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.441341 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-config-volume" (OuterVolumeSpecName: "config-volume") pod "9e1170b5-6a86-479d-9f29-dcffda7f5ba7" (UID: "9e1170b5-6a86-479d-9f29-dcffda7f5ba7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.447045 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-kube-api-access-fb2pd" (OuterVolumeSpecName: "kube-api-access-fb2pd") pod "9e1170b5-6a86-479d-9f29-dcffda7f5ba7" (UID: "9e1170b5-6a86-479d-9f29-dcffda7f5ba7"). InnerVolumeSpecName "kube-api-access-fb2pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.458080 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9e1170b5-6a86-479d-9f29-dcffda7f5ba7" (UID: "9e1170b5-6a86-479d-9f29-dcffda7f5ba7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.542231 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.542274 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb2pd\" (UniqueName: \"kubernetes.io/projected/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-kube-api-access-fb2pd\") on node \"crc\" DevicePath \"\"" Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.542283 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e1170b5-6a86-479d-9f29-dcffda7f5ba7-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.964263 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" event={"ID":"9e1170b5-6a86-479d-9f29-dcffda7f5ba7","Type":"ContainerDied","Data":"438f26f1c5c44d4244993201bb7a205d9979c7d36bfec1f450cafaf7860a6419"} Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.964970 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="438f26f1c5c44d4244993201bb7a205d9979c7d36bfec1f450cafaf7860a6419" Nov 24 08:00:03 crc kubenswrapper[4799]: I1124 08:00:03.964305 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc" Nov 24 08:00:04 crc kubenswrapper[4799]: I1124 08:00:04.430146 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c"] Nov 24 08:00:04 crc kubenswrapper[4799]: I1124 08:00:04.437105 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399475-rqf9c"] Nov 24 08:00:05 crc kubenswrapper[4799]: I1124 08:00:05.640195 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37" path="/var/lib/kubelet/pods/54ebc2b5-1913-488d-b6b5-e5c9e2ff2c37/volumes" Nov 24 08:00:07 crc kubenswrapper[4799]: I1124 08:00:07.630479 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:00:07 crc kubenswrapper[4799]: E1124 08:00:07.631314 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:00:19 crc kubenswrapper[4799]: I1124 08:00:19.630730 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:00:19 crc kubenswrapper[4799]: E1124 08:00:19.631696 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:00:31 crc kubenswrapper[4799]: I1124 08:00:31.628534 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:00:31 crc kubenswrapper[4799]: E1124 08:00:31.629593 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:00:45 crc kubenswrapper[4799]: I1124 08:00:45.635722 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:00:45 crc kubenswrapper[4799]: E1124 08:00:45.637342 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:00:54 crc kubenswrapper[4799]: I1124 08:00:54.321425 4799 scope.go:117] "RemoveContainer" containerID="f9daf8b045e1ae6d8fb7ef3e6f13b5546f3d083e0d85b8c6d6cd6003673aeee6" Nov 24 08:01:00 crc kubenswrapper[4799]: I1124 08:01:00.628145 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:01:00 crc kubenswrapper[4799]: E1124 08:01:00.629300 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:01:15 crc kubenswrapper[4799]: I1124 08:01:15.636558 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:01:15 crc kubenswrapper[4799]: E1124 08:01:15.637748 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:01:29 crc kubenswrapper[4799]: I1124 08:01:29.628035 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:01:29 crc kubenswrapper[4799]: E1124 08:01:29.628912 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:01:40 crc kubenswrapper[4799]: I1124 08:01:40.629059 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:01:40 crc kubenswrapper[4799]: E1124 08:01:40.630203 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:01:54 crc kubenswrapper[4799]: I1124 08:01:54.627741 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:01:54 crc kubenswrapper[4799]: E1124 08:01:54.628475 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:02:06 crc kubenswrapper[4799]: I1124 08:02:06.628341 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:02:06 crc kubenswrapper[4799]: E1124 08:02:06.629670 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:02:19 crc kubenswrapper[4799]: I1124 08:02:19.629144 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:02:19 crc kubenswrapper[4799]: E1124 08:02:19.630210 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.511471 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 08:02:20 crc kubenswrapper[4799]: E1124 08:02:20.512137 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e1170b5-6a86-479d-9f29-dcffda7f5ba7" containerName="collect-profiles" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.512162 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e1170b5-6a86-479d-9f29-dcffda7f5ba7" containerName="collect-profiles" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.512368 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e1170b5-6a86-479d-9f29-dcffda7f5ba7" containerName="collect-profiles" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.513057 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.517391 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zmjxb" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.540778 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.605126 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttjgk\" (UniqueName: \"kubernetes.io/projected/90f89a97-f122-4910-b4be-52bd61722372-kube-api-access-ttjgk\") pod \"mariadb-copy-data\" (UID: \"90f89a97-f122-4910-b4be-52bd61722372\") " pod="openstack/mariadb-copy-data" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.605215 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\") pod \"mariadb-copy-data\" (UID: \"90f89a97-f122-4910-b4be-52bd61722372\") " pod="openstack/mariadb-copy-data" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.706096 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\") pod \"mariadb-copy-data\" (UID: \"90f89a97-f122-4910-b4be-52bd61722372\") " pod="openstack/mariadb-copy-data" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.706685 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttjgk\" (UniqueName: \"kubernetes.io/projected/90f89a97-f122-4910-b4be-52bd61722372-kube-api-access-ttjgk\") pod \"mariadb-copy-data\" (UID: \"90f89a97-f122-4910-b4be-52bd61722372\") " pod="openstack/mariadb-copy-data" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.708964 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.709014 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\") pod \"mariadb-copy-data\" (UID: \"90f89a97-f122-4910-b4be-52bd61722372\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5621d7e839ea8ce7a8c2a5bbef76caa34437bdc32ac379edd5234c3ce81caa32/globalmount\"" pod="openstack/mariadb-copy-data" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.734660 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttjgk\" (UniqueName: \"kubernetes.io/projected/90f89a97-f122-4910-b4be-52bd61722372-kube-api-access-ttjgk\") pod \"mariadb-copy-data\" (UID: \"90f89a97-f122-4910-b4be-52bd61722372\") " pod="openstack/mariadb-copy-data" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.751118 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\") pod \"mariadb-copy-data\" (UID: \"90f89a97-f122-4910-b4be-52bd61722372\") " pod="openstack/mariadb-copy-data" Nov 24 08:02:20 crc kubenswrapper[4799]: I1124 08:02:20.839290 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 08:02:21 crc kubenswrapper[4799]: I1124 08:02:21.376917 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 08:02:22 crc kubenswrapper[4799]: I1124 08:02:22.374725 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"90f89a97-f122-4910-b4be-52bd61722372","Type":"ContainerStarted","Data":"fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb"} Nov 24 08:02:22 crc kubenswrapper[4799]: I1124 08:02:22.375234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"90f89a97-f122-4910-b4be-52bd61722372","Type":"ContainerStarted","Data":"9e07c6b90a00ae2d4041cce3a19a0b2314f5e4b84b3b3848c2fb5060f6bf2061"} Nov 24 08:02:22 crc kubenswrapper[4799]: I1124 08:02:22.401950 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.401920455 podStartE2EDuration="3.401920455s" podCreationTimestamp="2025-11-24 08:02:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:02:22.389875432 +0000 UTC m=+4488.045857906" watchObservedRunningTime="2025-11-24 08:02:22.401920455 +0000 UTC m=+4488.057902959" Nov 24 08:02:25 crc kubenswrapper[4799]: I1124 08:02:25.082469 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Nov 24 08:02:25 crc kubenswrapper[4799]: I1124 08:02:25.083772 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 08:02:25 crc kubenswrapper[4799]: I1124 08:02:25.094277 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 08:02:25 crc kubenswrapper[4799]: I1124 08:02:25.181951 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcjdg\" (UniqueName: \"kubernetes.io/projected/3122d359-0c04-43e1-8c82-878ccbcb29e8-kube-api-access-jcjdg\") pod \"mariadb-client\" (UID: \"3122d359-0c04-43e1-8c82-878ccbcb29e8\") " pod="openstack/mariadb-client" Nov 24 08:02:25 crc kubenswrapper[4799]: I1124 08:02:25.283587 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcjdg\" (UniqueName: \"kubernetes.io/projected/3122d359-0c04-43e1-8c82-878ccbcb29e8-kube-api-access-jcjdg\") pod \"mariadb-client\" (UID: \"3122d359-0c04-43e1-8c82-878ccbcb29e8\") " pod="openstack/mariadb-client" Nov 24 08:02:25 crc kubenswrapper[4799]: I1124 08:02:25.308224 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcjdg\" (UniqueName: \"kubernetes.io/projected/3122d359-0c04-43e1-8c82-878ccbcb29e8-kube-api-access-jcjdg\") pod \"mariadb-client\" (UID: \"3122d359-0c04-43e1-8c82-878ccbcb29e8\") " pod="openstack/mariadb-client" Nov 24 08:02:25 crc kubenswrapper[4799]: I1124 08:02:25.406069 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 08:02:25 crc kubenswrapper[4799]: I1124 08:02:25.932375 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 08:02:25 crc kubenswrapper[4799]: W1124 08:02:25.943133 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3122d359_0c04_43e1_8c82_878ccbcb29e8.slice/crio-6d9757d943ae43b4d05e5801a1085cc16dabb870c401ba4bbb14b40942dbeae0 WatchSource:0}: Error finding container 6d9757d943ae43b4d05e5801a1085cc16dabb870c401ba4bbb14b40942dbeae0: Status 404 returned error can't find the container with id 6d9757d943ae43b4d05e5801a1085cc16dabb870c401ba4bbb14b40942dbeae0 Nov 24 08:02:26 crc kubenswrapper[4799]: I1124 08:02:26.415147 4799 generic.go:334] "Generic (PLEG): container finished" podID="3122d359-0c04-43e1-8c82-878ccbcb29e8" containerID="a17fa8cb688753166d5579b8d68314f312d32ea8181ffd85e54a8a3fdf260df5" exitCode=0 Nov 24 08:02:26 crc kubenswrapper[4799]: I1124 08:02:26.415244 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"3122d359-0c04-43e1-8c82-878ccbcb29e8","Type":"ContainerDied","Data":"a17fa8cb688753166d5579b8d68314f312d32ea8181ffd85e54a8a3fdf260df5"} Nov 24 08:02:26 crc kubenswrapper[4799]: I1124 08:02:26.416152 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"3122d359-0c04-43e1-8c82-878ccbcb29e8","Type":"ContainerStarted","Data":"6d9757d943ae43b4d05e5801a1085cc16dabb870c401ba4bbb14b40942dbeae0"} Nov 24 08:02:27 crc kubenswrapper[4799]: I1124 08:02:27.754569 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 08:02:27 crc kubenswrapper[4799]: I1124 08:02:27.778017 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_3122d359-0c04-43e1-8c82-878ccbcb29e8/mariadb-client/0.log" Nov 24 08:02:27 crc kubenswrapper[4799]: I1124 08:02:27.811168 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Nov 24 08:02:27 crc kubenswrapper[4799]: I1124 08:02:27.815457 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Nov 24 08:02:27 crc kubenswrapper[4799]: I1124 08:02:27.831231 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcjdg\" (UniqueName: \"kubernetes.io/projected/3122d359-0c04-43e1-8c82-878ccbcb29e8-kube-api-access-jcjdg\") pod \"3122d359-0c04-43e1-8c82-878ccbcb29e8\" (UID: \"3122d359-0c04-43e1-8c82-878ccbcb29e8\") " Nov 24 08:02:27 crc kubenswrapper[4799]: I1124 08:02:27.949459 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Nov 24 08:02:27 crc kubenswrapper[4799]: E1124 08:02:27.949905 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3122d359-0c04-43e1-8c82-878ccbcb29e8" containerName="mariadb-client" Nov 24 08:02:27 crc kubenswrapper[4799]: I1124 08:02:27.949922 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3122d359-0c04-43e1-8c82-878ccbcb29e8" containerName="mariadb-client" Nov 24 08:02:27 crc kubenswrapper[4799]: I1124 08:02:27.950080 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3122d359-0c04-43e1-8c82-878ccbcb29e8" containerName="mariadb-client" Nov 24 08:02:27 crc kubenswrapper[4799]: I1124 08:02:27.950605 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 08:02:27 crc kubenswrapper[4799]: I1124 08:02:27.955079 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 08:02:28 crc kubenswrapper[4799]: I1124 08:02:28.035033 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqc25\" (UniqueName: \"kubernetes.io/projected/5f714c78-2fdb-4357-8088-b4cffb70d670-kube-api-access-pqc25\") pod \"mariadb-client\" (UID: \"5f714c78-2fdb-4357-8088-b4cffb70d670\") " pod="openstack/mariadb-client" Nov 24 08:02:28 crc kubenswrapper[4799]: I1124 08:02:28.137302 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqc25\" (UniqueName: \"kubernetes.io/projected/5f714c78-2fdb-4357-8088-b4cffb70d670-kube-api-access-pqc25\") pod \"mariadb-client\" (UID: \"5f714c78-2fdb-4357-8088-b4cffb70d670\") " pod="openstack/mariadb-client" Nov 24 08:02:28 crc kubenswrapper[4799]: I1124 08:02:28.271908 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3122d359-0c04-43e1-8c82-878ccbcb29e8-kube-api-access-jcjdg" (OuterVolumeSpecName: "kube-api-access-jcjdg") pod "3122d359-0c04-43e1-8c82-878ccbcb29e8" (UID: "3122d359-0c04-43e1-8c82-878ccbcb29e8"). InnerVolumeSpecName "kube-api-access-jcjdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:02:28 crc kubenswrapper[4799]: I1124 08:02:28.277384 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqc25\" (UniqueName: \"kubernetes.io/projected/5f714c78-2fdb-4357-8088-b4cffb70d670-kube-api-access-pqc25\") pod \"mariadb-client\" (UID: \"5f714c78-2fdb-4357-8088-b4cffb70d670\") " pod="openstack/mariadb-client" Nov 24 08:02:28 crc kubenswrapper[4799]: I1124 08:02:28.339981 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcjdg\" (UniqueName: \"kubernetes.io/projected/3122d359-0c04-43e1-8c82-878ccbcb29e8-kube-api-access-jcjdg\") on node \"crc\" DevicePath \"\"" Nov 24 08:02:28 crc kubenswrapper[4799]: I1124 08:02:28.435933 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d9757d943ae43b4d05e5801a1085cc16dabb870c401ba4bbb14b40942dbeae0" Nov 24 08:02:28 crc kubenswrapper[4799]: I1124 08:02:28.436356 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 08:02:28 crc kubenswrapper[4799]: I1124 08:02:28.467774 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="3122d359-0c04-43e1-8c82-878ccbcb29e8" podUID="5f714c78-2fdb-4357-8088-b4cffb70d670" Nov 24 08:02:28 crc kubenswrapper[4799]: I1124 08:02:28.564820 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 08:02:29 crc kubenswrapper[4799]: I1124 08:02:29.650450 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3122d359-0c04-43e1-8c82-878ccbcb29e8" path="/var/lib/kubelet/pods/3122d359-0c04-43e1-8c82-878ccbcb29e8/volumes" Nov 24 08:02:29 crc kubenswrapper[4799]: I1124 08:02:29.651803 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 08:02:30 crc kubenswrapper[4799]: W1124 08:02:30.074743 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f714c78_2fdb_4357_8088_b4cffb70d670.slice/crio-e5e7f5b50f6e4c94fa652626a399bfe2ff126601b2cf89e5c7fe2a98469fcbe7 WatchSource:0}: Error finding container e5e7f5b50f6e4c94fa652626a399bfe2ff126601b2cf89e5c7fe2a98469fcbe7: Status 404 returned error can't find the container with id e5e7f5b50f6e4c94fa652626a399bfe2ff126601b2cf89e5c7fe2a98469fcbe7 Nov 24 08:02:30 crc kubenswrapper[4799]: I1124 08:02:30.461426 4799 generic.go:334] "Generic (PLEG): container finished" podID="5f714c78-2fdb-4357-8088-b4cffb70d670" containerID="36af58232f124ebd95ee1c8c030dcf54d41894292ac10278c731360025a59ed8" exitCode=0 Nov 24 08:02:30 crc kubenswrapper[4799]: I1124 08:02:30.461511 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"5f714c78-2fdb-4357-8088-b4cffb70d670","Type":"ContainerDied","Data":"36af58232f124ebd95ee1c8c030dcf54d41894292ac10278c731360025a59ed8"} Nov 24 08:02:30 crc kubenswrapper[4799]: I1124 08:02:30.461563 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"5f714c78-2fdb-4357-8088-b4cffb70d670","Type":"ContainerStarted","Data":"e5e7f5b50f6e4c94fa652626a399bfe2ff126601b2cf89e5c7fe2a98469fcbe7"} Nov 24 08:02:31 crc kubenswrapper[4799]: I1124 08:02:31.629933 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:02:31 crc kubenswrapper[4799]: E1124 08:02:31.631305 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:02:31 crc kubenswrapper[4799]: I1124 08:02:31.795784 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 08:02:31 crc kubenswrapper[4799]: I1124 08:02:31.815279 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_5f714c78-2fdb-4357-8088-b4cffb70d670/mariadb-client/0.log" Nov 24 08:02:31 crc kubenswrapper[4799]: I1124 08:02:31.844955 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Nov 24 08:02:31 crc kubenswrapper[4799]: I1124 08:02:31.851978 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Nov 24 08:02:31 crc kubenswrapper[4799]: I1124 08:02:31.896714 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqc25\" (UniqueName: \"kubernetes.io/projected/5f714c78-2fdb-4357-8088-b4cffb70d670-kube-api-access-pqc25\") pod \"5f714c78-2fdb-4357-8088-b4cffb70d670\" (UID: \"5f714c78-2fdb-4357-8088-b4cffb70d670\") " Nov 24 08:02:31 crc kubenswrapper[4799]: I1124 08:02:31.901771 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f714c78-2fdb-4357-8088-b4cffb70d670-kube-api-access-pqc25" (OuterVolumeSpecName: "kube-api-access-pqc25") pod "5f714c78-2fdb-4357-8088-b4cffb70d670" (UID: "5f714c78-2fdb-4357-8088-b4cffb70d670"). InnerVolumeSpecName "kube-api-access-pqc25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:02:31 crc kubenswrapper[4799]: I1124 08:02:31.998914 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqc25\" (UniqueName: \"kubernetes.io/projected/5f714c78-2fdb-4357-8088-b4cffb70d670-kube-api-access-pqc25\") on node \"crc\" DevicePath \"\"" Nov 24 08:02:32 crc kubenswrapper[4799]: I1124 08:02:32.480552 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5e7f5b50f6e4c94fa652626a399bfe2ff126601b2cf89e5c7fe2a98469fcbe7" Nov 24 08:02:32 crc kubenswrapper[4799]: I1124 08:02:32.480685 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 08:02:33 crc kubenswrapper[4799]: I1124 08:02:33.640783 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f714c78-2fdb-4357-8088-b4cffb70d670" path="/var/lib/kubelet/pods/5f714c78-2fdb-4357-8088-b4cffb70d670/volumes" Nov 24 08:02:45 crc kubenswrapper[4799]: I1124 08:02:45.634170 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:02:45 crc kubenswrapper[4799]: E1124 08:02:45.635005 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:03:00 crc kubenswrapper[4799]: I1124 08:03:00.628543 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:03:00 crc kubenswrapper[4799]: E1124 08:03:00.629462 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.804181 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 08:03:03 crc kubenswrapper[4799]: E1124 08:03:03.806658 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f714c78-2fdb-4357-8088-b4cffb70d670" containerName="mariadb-client" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.806809 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f714c78-2fdb-4357-8088-b4cffb70d670" containerName="mariadb-client" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.807378 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f714c78-2fdb-4357-8088-b4cffb70d670" containerName="mariadb-client" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.808989 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.816419 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.817164 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.817925 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-plqsj" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.837399 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.845671 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.850102 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.855399 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.857443 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.867136 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmvr2\" (UniqueName: \"kubernetes.io/projected/2b5580f4-5a36-492a-a9de-cdae05bda4a7-kube-api-access-vmvr2\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.867199 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b5580f4-5a36-492a-a9de-cdae05bda4a7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.867227 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b5580f4-5a36-492a-a9de-cdae05bda4a7-config\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.867255 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b5580f4-5a36-492a-a9de-cdae05bda4a7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.867295 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0cb40107-57ce-4849-8a0c-e0d43d0e636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0cb40107-57ce-4849-8a0c-e0d43d0e636d\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.867317 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2b5580f4-5a36-492a-a9de-cdae05bda4a7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.878175 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.887412 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.971967 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0cb40107-57ce-4849-8a0c-e0d43d0e636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0cb40107-57ce-4849-8a0c-e0d43d0e636d\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972012 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2b5580f4-5a36-492a-a9de-cdae05bda4a7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972049 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk48z\" (UniqueName: \"kubernetes.io/projected/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-kube-api-access-zk48z\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972070 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-69405f85-7ca7-4978-b995-fd2dc7230abe\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69405f85-7ca7-4978-b995-fd2dc7230abe\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972088 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/758e3db3-eb2f-48bf-8434-39d0bf4b8595-config\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972122 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-config\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972144 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758e3db3-eb2f-48bf-8434-39d0bf4b8595-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7a6ced77-8b99-4471-92c1-7785565784a0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7a6ced77-8b99-4471-92c1-7785565784a0\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972184 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972200 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/758e3db3-eb2f-48bf-8434-39d0bf4b8595-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972229 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmvr2\" (UniqueName: \"kubernetes.io/projected/2b5580f4-5a36-492a-a9de-cdae05bda4a7-kube-api-access-vmvr2\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972252 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972269 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnprh\" (UniqueName: \"kubernetes.io/projected/758e3db3-eb2f-48bf-8434-39d0bf4b8595-kube-api-access-fnprh\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972288 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b5580f4-5a36-492a-a9de-cdae05bda4a7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972307 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b5580f4-5a36-492a-a9de-cdae05bda4a7-config\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972327 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b5580f4-5a36-492a-a9de-cdae05bda4a7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972344 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972362 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/758e3db3-eb2f-48bf-8434-39d0bf4b8595-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.972490 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2b5580f4-5a36-492a-a9de-cdae05bda4a7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.974112 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b5580f4-5a36-492a-a9de-cdae05bda4a7-config\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.974557 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b5580f4-5a36-492a-a9de-cdae05bda4a7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.975667 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.975770 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0cb40107-57ce-4849-8a0c-e0d43d0e636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0cb40107-57ce-4849-8a0c-e0d43d0e636d\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9db1faeba05f2ca4bb1b6634ed6b6686ae6fbcd3500621a0347c9d2361d79be2/globalmount\"" pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.985182 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b5580f4-5a36-492a-a9de-cdae05bda4a7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.988027 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmvr2\" (UniqueName: \"kubernetes.io/projected/2b5580f4-5a36-492a-a9de-cdae05bda4a7-kube-api-access-vmvr2\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:03 crc kubenswrapper[4799]: I1124 08:03:03.999363 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.002133 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.007717 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.007921 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.008183 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-8cksn" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.015088 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.030252 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0cb40107-57ce-4849-8a0c-e0d43d0e636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0cb40107-57ce-4849-8a0c-e0d43d0e636d\") pod \"ovsdbserver-nb-0\" (UID: \"2b5580f4-5a36-492a-a9de-cdae05bda4a7\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.034398 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.035794 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.043959 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.045558 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.055989 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.062040 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.072963 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5703958e-9f91-479c-a89b-c90494eb07ec-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.072995 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c136e238-389c-490f-a6f5-afa0f00c55dd-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073024 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073043 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/758e3db3-eb2f-48bf-8434-39d0bf4b8595-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073062 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e03fcece-6fcb-4fe1-9199-67819ee430e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e03fcece-6fcb-4fe1-9199-67819ee430e1\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073077 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c136e238-389c-490f-a6f5-afa0f00c55dd-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073091 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5703958e-9f91-479c-a89b-c90494eb07ec-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073119 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073136 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073155 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnprh\" (UniqueName: \"kubernetes.io/projected/758e3db3-eb2f-48bf-8434-39d0bf4b8595-kube-api-access-fnprh\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073181 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c136e238-389c-490f-a6f5-afa0f00c55dd-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073197 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdwx4\" (UniqueName: \"kubernetes.io/projected/5703958e-9f91-479c-a89b-c90494eb07ec-kube-api-access-rdwx4\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073221 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073237 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073256 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6dc4e2ac-8cb2-4182-b546-2dabce3bbbf0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6dc4e2ac-8cb2-4182-b546-2dabce3bbbf0\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073274 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/758e3db3-eb2f-48bf-8434-39d0bf4b8595-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073291 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c136e238-389c-490f-a6f5-afa0f00c55dd-config\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073319 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-config\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073337 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk48z\" (UniqueName: \"kubernetes.io/projected/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-kube-api-access-zk48z\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073354 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-69405f85-7ca7-4978-b995-fd2dc7230abe\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69405f85-7ca7-4978-b995-fd2dc7230abe\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073369 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/758e3db3-eb2f-48bf-8434-39d0bf4b8595-config\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073383 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5703958e-9f91-479c-a89b-c90494eb07ec-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073403 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-99924099-94c7-42fb-bc80-cd5128c79098\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-99924099-94c7-42fb-bc80-cd5128c79098\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073426 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksp7z\" (UniqueName: \"kubernetes.io/projected/c136e238-389c-490f-a6f5-afa0f00c55dd-kube-api-access-ksp7z\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073434 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/758e3db3-eb2f-48bf-8434-39d0bf4b8595-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073443 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtfrr\" (UniqueName: \"kubernetes.io/projected/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-kube-api-access-gtfrr\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073483 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-config\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073506 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073525 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5703958e-9f91-479c-a89b-c90494eb07ec-config\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073542 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758e3db3-eb2f-48bf-8434-39d0bf4b8595-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073564 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7a6ced77-8b99-4471-92c1-7785565784a0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7a6ced77-8b99-4471-92c1-7785565784a0\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.073889 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.074370 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-config\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.074808 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.075144 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/758e3db3-eb2f-48bf-8434-39d0bf4b8595-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.075357 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/758e3db3-eb2f-48bf-8434-39d0bf4b8595-config\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.076261 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.076292 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-69405f85-7ca7-4978-b995-fd2dc7230abe\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69405f85-7ca7-4978-b995-fd2dc7230abe\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/aef83eb940c038fbdadb40951f8825699da6725d8193d92d6565da1e3cef4bb5/globalmount\"" pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.076335 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.076359 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7a6ced77-8b99-4471-92c1-7785565784a0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7a6ced77-8b99-4471-92c1-7785565784a0\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/844ea8f03934c4f4c660e41409f166b6dabe6ea262b1ca2889b85766463fef2e/globalmount\"" pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.077253 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.078282 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758e3db3-eb2f-48bf-8434-39d0bf4b8595-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.090295 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk48z\" (UniqueName: \"kubernetes.io/projected/f99fe0ba-86b1-47ce-b936-c40ab7e4539e-kube-api-access-zk48z\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.091656 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnprh\" (UniqueName: \"kubernetes.io/projected/758e3db3-eb2f-48bf-8434-39d0bf4b8595-kube-api-access-fnprh\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.107710 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-69405f85-7ca7-4978-b995-fd2dc7230abe\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69405f85-7ca7-4978-b995-fd2dc7230abe\") pod \"ovsdbserver-nb-2\" (UID: \"758e3db3-eb2f-48bf-8434-39d0bf4b8595\") " pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.109185 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7a6ced77-8b99-4471-92c1-7785565784a0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7a6ced77-8b99-4471-92c1-7785565784a0\") pod \"ovsdbserver-nb-1\" (UID: \"f99fe0ba-86b1-47ce-b936-c40ab7e4539e\") " pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.141658 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174569 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174619 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c136e238-389c-490f-a6f5-afa0f00c55dd-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174640 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwx4\" (UniqueName: \"kubernetes.io/projected/5703958e-9f91-479c-a89b-c90494eb07ec-kube-api-access-rdwx4\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174667 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174688 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6dc4e2ac-8cb2-4182-b546-2dabce3bbbf0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6dc4e2ac-8cb2-4182-b546-2dabce3bbbf0\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174706 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c136e238-389c-490f-a6f5-afa0f00c55dd-config\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174735 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-config\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174758 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5703958e-9f91-479c-a89b-c90494eb07ec-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174779 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-99924099-94c7-42fb-bc80-cd5128c79098\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-99924099-94c7-42fb-bc80-cd5128c79098\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174808 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksp7z\" (UniqueName: \"kubernetes.io/projected/c136e238-389c-490f-a6f5-afa0f00c55dd-kube-api-access-ksp7z\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174827 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtfrr\" (UniqueName: \"kubernetes.io/projected/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-kube-api-access-gtfrr\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174871 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174891 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5703958e-9f91-479c-a89b-c90494eb07ec-config\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174917 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5703958e-9f91-479c-a89b-c90494eb07ec-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.174975 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c136e238-389c-490f-a6f5-afa0f00c55dd-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.175001 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e03fcece-6fcb-4fe1-9199-67819ee430e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e03fcece-6fcb-4fe1-9199-67819ee430e1\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.175049 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c136e238-389c-490f-a6f5-afa0f00c55dd-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.175075 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5703958e-9f91-479c-a89b-c90494eb07ec-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.176728 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c136e238-389c-490f-a6f5-afa0f00c55dd-config\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.177233 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.177303 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5703958e-9f91-479c-a89b-c90494eb07ec-config\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.178045 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c136e238-389c-490f-a6f5-afa0f00c55dd-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.178490 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5703958e-9f91-479c-a89b-c90494eb07ec-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.178952 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c136e238-389c-490f-a6f5-afa0f00c55dd-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.179366 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5703958e-9f91-479c-a89b-c90494eb07ec-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.180402 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5703958e-9f91-479c-a89b-c90494eb07ec-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.180698 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.180809 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e03fcece-6fcb-4fe1-9199-67819ee430e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e03fcece-6fcb-4fe1-9199-67819ee430e1\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2661debbaee84d60b18f702f77fed49fd115a4e8eda187f0222f84970bcf5833/globalmount\"" pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.181562 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.181863 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.181892 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-99924099-94c7-42fb-bc80-cd5128c79098\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-99924099-94c7-42fb-bc80-cd5128c79098\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8afc972e6bdb2ace6a89907cf72b003bc29bd9f035d6e11b649e365187d85dc2/globalmount\"" pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.181892 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c136e238-389c-490f-a6f5-afa0f00c55dd-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.182442 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.182539 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6dc4e2ac-8cb2-4182-b546-2dabce3bbbf0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6dc4e2ac-8cb2-4182-b546-2dabce3bbbf0\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0b922e78d7d8d9e45a8b5a35fc7fd59d2b499b3c301deceea7f647e8c1ea8122/globalmount\"" pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.186071 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.191339 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.195087 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtfrr\" (UniqueName: \"kubernetes.io/projected/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-kube-api-access-gtfrr\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.197061 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ef83552-6d16-407f-a4ed-bd0a8f9c5927-config\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.201405 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksp7z\" (UniqueName: \"kubernetes.io/projected/c136e238-389c-490f-a6f5-afa0f00c55dd-kube-api-access-ksp7z\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.201409 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwx4\" (UniqueName: \"kubernetes.io/projected/5703958e-9f91-479c-a89b-c90494eb07ec-kube-api-access-rdwx4\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.203321 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.230807 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-99924099-94c7-42fb-bc80-cd5128c79098\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-99924099-94c7-42fb-bc80-cd5128c79098\") pod \"ovsdbserver-sb-0\" (UID: \"0ef83552-6d16-407f-a4ed-bd0a8f9c5927\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.235634 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e03fcece-6fcb-4fe1-9199-67819ee430e1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e03fcece-6fcb-4fe1-9199-67819ee430e1\") pod \"ovsdbserver-sb-2\" (UID: \"c136e238-389c-490f-a6f5-afa0f00c55dd\") " pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.248359 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6dc4e2ac-8cb2-4182-b546-2dabce3bbbf0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6dc4e2ac-8cb2-4182-b546-2dabce3bbbf0\") pod \"ovsdbserver-sb-1\" (UID: \"5703958e-9f91-479c-a89b-c90494eb07ec\") " pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.376217 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.451331 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.492425 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.658867 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.669946 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.763383 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.831618 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"758e3db3-eb2f-48bf-8434-39d0bf4b8595","Type":"ContainerStarted","Data":"56eca50c8fe43e997ae9ac598df13cf511503f9034a0a5eda2eead5e3e298238"} Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.833350 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2b5580f4-5a36-492a-a9de-cdae05bda4a7","Type":"ContainerStarted","Data":"004793927cf152b9ce0beab8614b2d67abc72ee269b629a304fae3b7ea1c34ee"} Nov 24 08:03:04 crc kubenswrapper[4799]: W1124 08:03:04.859839 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf99fe0ba_86b1_47ce_b936_c40ab7e4539e.slice/crio-45f0d6873c4dbcf05cc9ab8b9ca17a5a72ff0df9743242d9cd3d2f30c463d1df WatchSource:0}: Error finding container 45f0d6873c4dbcf05cc9ab8b9ca17a5a72ff0df9743242d9cd3d2f30c463d1df: Status 404 returned error can't find the container with id 45f0d6873c4dbcf05cc9ab8b9ca17a5a72ff0df9743242d9cd3d2f30c463d1df Nov 24 08:03:04 crc kubenswrapper[4799]: I1124 08:03:04.859952 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 24 08:03:05 crc kubenswrapper[4799]: I1124 08:03:05.068982 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 24 08:03:05 crc kubenswrapper[4799]: W1124 08:03:05.072308 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5703958e_9f91_479c_a89b_c90494eb07ec.slice/crio-7b407ab24174010dbcbdeb3fb27aa6e25c6a458e13d83cbf4a29085533dab912 WatchSource:0}: Error finding container 7b407ab24174010dbcbdeb3fb27aa6e25c6a458e13d83cbf4a29085533dab912: Status 404 returned error can't find the container with id 7b407ab24174010dbcbdeb3fb27aa6e25c6a458e13d83cbf4a29085533dab912 Nov 24 08:03:05 crc kubenswrapper[4799]: I1124 08:03:05.581227 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 08:03:05 crc kubenswrapper[4799]: W1124 08:03:05.590724 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ef83552_6d16_407f_a4ed_bd0a8f9c5927.slice/crio-293e7ed9c780581772e661b8b40fc7feb67dff587230a5d75db6835b1b515238 WatchSource:0}: Error finding container 293e7ed9c780581772e661b8b40fc7feb67dff587230a5d75db6835b1b515238: Status 404 returned error can't find the container with id 293e7ed9c780581772e661b8b40fc7feb67dff587230a5d75db6835b1b515238 Nov 24 08:03:05 crc kubenswrapper[4799]: I1124 08:03:05.705378 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 24 08:03:05 crc kubenswrapper[4799]: I1124 08:03:05.843012 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0ef83552-6d16-407f-a4ed-bd0a8f9c5927","Type":"ContainerStarted","Data":"293e7ed9c780581772e661b8b40fc7feb67dff587230a5d75db6835b1b515238"} Nov 24 08:03:05 crc kubenswrapper[4799]: I1124 08:03:05.844822 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"5703958e-9f91-479c-a89b-c90494eb07ec","Type":"ContainerStarted","Data":"7b407ab24174010dbcbdeb3fb27aa6e25c6a458e13d83cbf4a29085533dab912"} Nov 24 08:03:05 crc kubenswrapper[4799]: I1124 08:03:05.846521 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"f99fe0ba-86b1-47ce-b936-c40ab7e4539e","Type":"ContainerStarted","Data":"45f0d6873c4dbcf05cc9ab8b9ca17a5a72ff0df9743242d9cd3d2f30c463d1df"} Nov 24 08:03:05 crc kubenswrapper[4799]: I1124 08:03:05.848495 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"c136e238-389c-490f-a6f5-afa0f00c55dd","Type":"ContainerStarted","Data":"c89cd21113a3f3b4f434c7c038f2eaa0ae125105a0e1fc87f15a8bc2ec48a5c4"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.883608 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"f99fe0ba-86b1-47ce-b936-c40ab7e4539e","Type":"ContainerStarted","Data":"6aee10520c66ea11965b4812e90f7949b0541d10b9a19e1f91b12db321ff2e85"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.884737 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"f99fe0ba-86b1-47ce-b936-c40ab7e4539e","Type":"ContainerStarted","Data":"4156f885b9711c3f6e66ebd410275dadee29e5d07ea17522f391f064fe4af381"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.891267 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"c136e238-389c-490f-a6f5-afa0f00c55dd","Type":"ContainerStarted","Data":"476463e8c3c3344810f1d9b5ec98d2cb962455b8fad85e2c1105dfa3716a30ee"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.891319 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"c136e238-389c-490f-a6f5-afa0f00c55dd","Type":"ContainerStarted","Data":"bcbd07dd82473228dd3e605bcbf27cb2b74312e6544766373fd805b0343ee069"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.894211 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2b5580f4-5a36-492a-a9de-cdae05bda4a7","Type":"ContainerStarted","Data":"078bfece9e3a25d7e9504ffaae19d851161da73d678d5aa85b319a6feb047e8e"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.894283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2b5580f4-5a36-492a-a9de-cdae05bda4a7","Type":"ContainerStarted","Data":"1b5460c88b9d87dd32da1463414cfaa380b1235b7bfcef15c88cb18d6b7b1dea"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.895679 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0ef83552-6d16-407f-a4ed-bd0a8f9c5927","Type":"ContainerStarted","Data":"77574c9bbc40124101296e327b9c7a58da36ad77ada5f317c99c71f1b5164b13"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.897875 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"5703958e-9f91-479c-a89b-c90494eb07ec","Type":"ContainerStarted","Data":"9902f8e37581654c18360a95b6cd3ba872957dcfcc55fbac804b2100de52ba2f"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.897911 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"5703958e-9f91-479c-a89b-c90494eb07ec","Type":"ContainerStarted","Data":"67e633d84d8118efb60cf54f03f49cb38e0134c0f0cbd9f83224018d9ddd1ca7"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.903762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"758e3db3-eb2f-48bf-8434-39d0bf4b8595","Type":"ContainerStarted","Data":"72d830c7fe93ec626904b6bb94f89bb9990341314a15c137b9d6f3df5fab861c"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.903791 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"758e3db3-eb2f-48bf-8434-39d0bf4b8595","Type":"ContainerStarted","Data":"ae24ea6cb669d8ca4ffcb27824832dd9215b37e6b071db1287a0d87e6b900283"} Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.919336 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.9972305820000003 podStartE2EDuration="7.919298103s" podCreationTimestamp="2025-11-24 08:03:02 +0000 UTC" firstStartedPulling="2025-11-24 08:03:04.862457431 +0000 UTC m=+4530.518439905" lastFinishedPulling="2025-11-24 08:03:08.784524952 +0000 UTC m=+4534.440507426" observedRunningTime="2025-11-24 08:03:09.911913463 +0000 UTC m=+4535.567895957" watchObservedRunningTime="2025-11-24 08:03:09.919298103 +0000 UTC m=+4535.575280617" Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.947936 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=4.841413187 podStartE2EDuration="7.947906756s" podCreationTimestamp="2025-11-24 08:03:02 +0000 UTC" firstStartedPulling="2025-11-24 08:03:05.710241368 +0000 UTC m=+4531.366223842" lastFinishedPulling="2025-11-24 08:03:08.816734937 +0000 UTC m=+4534.472717411" observedRunningTime="2025-11-24 08:03:09.938469998 +0000 UTC m=+4535.594452462" watchObservedRunningTime="2025-11-24 08:03:09.947906756 +0000 UTC m=+4535.603889260" Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.967446 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.8416105 podStartE2EDuration="7.96742103s" podCreationTimestamp="2025-11-24 08:03:02 +0000 UTC" firstStartedPulling="2025-11-24 08:03:04.669663163 +0000 UTC m=+4530.325645647" lastFinishedPulling="2025-11-24 08:03:08.795473703 +0000 UTC m=+4534.451456177" observedRunningTime="2025-11-24 08:03:09.958462206 +0000 UTC m=+4535.614444680" watchObservedRunningTime="2025-11-24 08:03:09.96742103 +0000 UTC m=+4535.623403524" Nov 24 08:03:09 crc kubenswrapper[4799]: I1124 08:03:09.984971 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.975934807 podStartE2EDuration="7.984955668s" podCreationTimestamp="2025-11-24 08:03:02 +0000 UTC" firstStartedPulling="2025-11-24 08:03:04.767130963 +0000 UTC m=+4530.423113437" lastFinishedPulling="2025-11-24 08:03:08.776151814 +0000 UTC m=+4534.432134298" observedRunningTime="2025-11-24 08:03:09.981540701 +0000 UTC m=+4535.637523175" watchObservedRunningTime="2025-11-24 08:03:09.984955668 +0000 UTC m=+4535.640938142" Nov 24 08:03:10 crc kubenswrapper[4799]: I1124 08:03:10.004355 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=3.955291991 podStartE2EDuration="8.004333589s" podCreationTimestamp="2025-11-24 08:03:02 +0000 UTC" firstStartedPulling="2025-11-24 08:03:05.074366762 +0000 UTC m=+4530.730349236" lastFinishedPulling="2025-11-24 08:03:09.12340836 +0000 UTC m=+4534.779390834" observedRunningTime="2025-11-24 08:03:10.001258051 +0000 UTC m=+4535.657240515" watchObservedRunningTime="2025-11-24 08:03:10.004333589 +0000 UTC m=+4535.660316083" Nov 24 08:03:10 crc kubenswrapper[4799]: I1124 08:03:10.142076 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:10 crc kubenswrapper[4799]: I1124 08:03:10.192307 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:10 crc kubenswrapper[4799]: I1124 08:03:10.203994 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:10 crc kubenswrapper[4799]: I1124 08:03:10.452177 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:10 crc kubenswrapper[4799]: I1124 08:03:10.493233 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:10 crc kubenswrapper[4799]: I1124 08:03:10.915079 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0ef83552-6d16-407f-a4ed-bd0a8f9c5927","Type":"ContainerStarted","Data":"5e502c090f48733e2c1eb10d31b1d7dcc335753040aaf90f434c5d3a820f425f"} Nov 24 08:03:10 crc kubenswrapper[4799]: I1124 08:03:10.943962 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=5.194678633 podStartE2EDuration="8.943828542s" podCreationTimestamp="2025-11-24 08:03:02 +0000 UTC" firstStartedPulling="2025-11-24 08:03:05.594205571 +0000 UTC m=+4531.250188045" lastFinishedPulling="2025-11-24 08:03:09.34335548 +0000 UTC m=+4534.999337954" observedRunningTime="2025-11-24 08:03:10.939332124 +0000 UTC m=+4536.595314598" watchObservedRunningTime="2025-11-24 08:03:10.943828542 +0000 UTC m=+4536.599811056" Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.588287 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f5n4k"] Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.592698 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.604942 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5n4k"] Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.730294 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-utilities\") pod \"redhat-marketplace-f5n4k\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.730347 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdzt7\" (UniqueName: \"kubernetes.io/projected/1d7f6fcb-e604-4d23-b352-5de0f857f449-kube-api-access-kdzt7\") pod \"redhat-marketplace-f5n4k\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.730429 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-catalog-content\") pod \"redhat-marketplace-f5n4k\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.831374 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-utilities\") pod \"redhat-marketplace-f5n4k\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.831753 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdzt7\" (UniqueName: \"kubernetes.io/projected/1d7f6fcb-e604-4d23-b352-5de0f857f449-kube-api-access-kdzt7\") pod \"redhat-marketplace-f5n4k\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.831820 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-catalog-content\") pod \"redhat-marketplace-f5n4k\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.832177 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-utilities\") pod \"redhat-marketplace-f5n4k\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.832332 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-catalog-content\") pod \"redhat-marketplace-f5n4k\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:11 crc kubenswrapper[4799]: I1124 08:03:11.975997 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdzt7\" (UniqueName: \"kubernetes.io/projected/1d7f6fcb-e604-4d23-b352-5de0f857f449-kube-api-access-kdzt7\") pod \"redhat-marketplace-f5n4k\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:12 crc kubenswrapper[4799]: I1124 08:03:12.224243 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:12 crc kubenswrapper[4799]: I1124 08:03:12.669457 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5n4k"] Nov 24 08:03:12 crc kubenswrapper[4799]: W1124 08:03:12.672220 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d7f6fcb_e604_4d23_b352_5de0f857f449.slice/crio-a25902e38d3763686abb16929a03d91fc82790089b458b08f818e62cd2a7a998 WatchSource:0}: Error finding container a25902e38d3763686abb16929a03d91fc82790089b458b08f818e62cd2a7a998: Status 404 returned error can't find the container with id a25902e38d3763686abb16929a03d91fc82790089b458b08f818e62cd2a7a998 Nov 24 08:03:12 crc kubenswrapper[4799]: I1124 08:03:12.933554 4799 generic.go:334] "Generic (PLEG): container finished" podID="1d7f6fcb-e604-4d23-b352-5de0f857f449" containerID="9d595b05b9e344365250989ce0d65ae815390f7e79aaa708520df489457fe089" exitCode=0 Nov 24 08:03:12 crc kubenswrapper[4799]: I1124 08:03:12.933606 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5n4k" event={"ID":"1d7f6fcb-e604-4d23-b352-5de0f857f449","Type":"ContainerDied","Data":"9d595b05b9e344365250989ce0d65ae815390f7e79aaa708520df489457fe089"} Nov 24 08:03:12 crc kubenswrapper[4799]: I1124 08:03:12.933637 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5n4k" event={"ID":"1d7f6fcb-e604-4d23-b352-5de0f857f449","Type":"ContainerStarted","Data":"a25902e38d3763686abb16929a03d91fc82790089b458b08f818e62cd2a7a998"} Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.226743 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.227393 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.249261 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.252821 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.258723 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.259298 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.377673 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.429520 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.492527 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.493311 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.556539 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.557092 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.951050 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5n4k" event={"ID":"1d7f6fcb-e604-4d23-b352-5de0f857f449","Type":"ContainerStarted","Data":"77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910"} Nov 24 08:03:13 crc kubenswrapper[4799]: I1124 08:03:13.952462 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.180660 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.243772 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.244419 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.407660 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57b9ddc79c-884r9"] Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.408957 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.412121 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.433694 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57b9ddc79c-884r9"] Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.476451 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-dns-svc\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.476501 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-ovsdbserver-nb\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.476569 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-config\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.476587 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxtbc\" (UniqueName: \"kubernetes.io/projected/1905c846-150f-488c-9f99-36d2f4bb985d-kube-api-access-wxtbc\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.552873 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.577822 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-dns-svc\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.577895 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-ovsdbserver-nb\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.577966 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-config\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.578002 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxtbc\" (UniqueName: \"kubernetes.io/projected/1905c846-150f-488c-9f99-36d2f4bb985d-kube-api-access-wxtbc\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.578746 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-dns-svc\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.579138 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-ovsdbserver-nb\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.579335 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-config\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.607878 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.629644 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.630183 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxtbc\" (UniqueName: \"kubernetes.io/projected/1905c846-150f-488c-9f99-36d2f4bb985d-kube-api-access-wxtbc\") pod \"dnsmasq-dns-57b9ddc79c-884r9\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.765579 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.941467 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57b9ddc79c-884r9"] Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.961969 4799 generic.go:334] "Generic (PLEG): container finished" podID="1d7f6fcb-e604-4d23-b352-5de0f857f449" containerID="77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910" exitCode=0 Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.962037 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5n4k" event={"ID":"1d7f6fcb-e604-4d23-b352-5de0f857f449","Type":"ContainerDied","Data":"77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910"} Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.970180 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59bcf85c7c-w569x"] Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.971917 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.979345 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 08:03:14 crc kubenswrapper[4799]: I1124 08:03:14.990054 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59bcf85c7c-w569x"] Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.088919 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7ccj\" (UniqueName: \"kubernetes.io/projected/80b671da-67db-4381-b646-d394eb8b43a0-kube-api-access-g7ccj\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.088978 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-sb\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.088998 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-dns-svc\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.089187 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-config\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.089261 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-nb\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.190868 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-config\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.190958 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-nb\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.191028 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7ccj\" (UniqueName: \"kubernetes.io/projected/80b671da-67db-4381-b646-d394eb8b43a0-kube-api-access-g7ccj\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.191061 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-sb\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.191085 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-dns-svc\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.191727 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-config\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.191914 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-nb\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.192020 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-dns-svc\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.192142 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-sb\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.222952 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7ccj\" (UniqueName: \"kubernetes.io/projected/80b671da-67db-4381-b646-d394eb8b43a0-kube-api-access-g7ccj\") pod \"dnsmasq-dns-59bcf85c7c-w569x\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.255216 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57b9ddc79c-884r9"] Nov 24 08:03:15 crc kubenswrapper[4799]: W1124 08:03:15.263246 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1905c846_150f_488c_9f99_36d2f4bb985d.slice/crio-a8c2ca0b6252aa21a69d3dbba358dc369c9f3a0230a23f1fda745cef3ae54974 WatchSource:0}: Error finding container a8c2ca0b6252aa21a69d3dbba358dc369c9f3a0230a23f1fda745cef3ae54974: Status 404 returned error can't find the container with id a8c2ca0b6252aa21a69d3dbba358dc369c9f3a0230a23f1fda745cef3ae54974 Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.308497 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.633443 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:03:15 crc kubenswrapper[4799]: E1124 08:03:15.633983 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.759604 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59bcf85c7c-w569x"] Nov 24 08:03:15 crc kubenswrapper[4799]: W1124 08:03:15.760739 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80b671da_67db_4381_b646_d394eb8b43a0.slice/crio-3c8353a616628d1136ed0e6207e88b9283b85258f8f53ccd6cb6aa98057efaff WatchSource:0}: Error finding container 3c8353a616628d1136ed0e6207e88b9283b85258f8f53ccd6cb6aa98057efaff: Status 404 returned error can't find the container with id 3c8353a616628d1136ed0e6207e88b9283b85258f8f53ccd6cb6aa98057efaff Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.970800 4799 generic.go:334] "Generic (PLEG): container finished" podID="1905c846-150f-488c-9f99-36d2f4bb985d" containerID="5f97ff01fe1764842f25b07bfd72bc52e73c63a23a565f154797fcd3b0e6cb6b" exitCode=0 Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.970875 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" event={"ID":"1905c846-150f-488c-9f99-36d2f4bb985d","Type":"ContainerDied","Data":"5f97ff01fe1764842f25b07bfd72bc52e73c63a23a565f154797fcd3b0e6cb6b"} Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.970930 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" event={"ID":"1905c846-150f-488c-9f99-36d2f4bb985d","Type":"ContainerStarted","Data":"a8c2ca0b6252aa21a69d3dbba358dc369c9f3a0230a23f1fda745cef3ae54974"} Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.974470 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5n4k" event={"ID":"1d7f6fcb-e604-4d23-b352-5de0f857f449","Type":"ContainerStarted","Data":"27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0"} Nov 24 08:03:15 crc kubenswrapper[4799]: I1124 08:03:15.980582 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" event={"ID":"80b671da-67db-4381-b646-d394eb8b43a0","Type":"ContainerStarted","Data":"3c8353a616628d1136ed0e6207e88b9283b85258f8f53ccd6cb6aa98057efaff"} Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.029763 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f5n4k" podStartSLOduration=2.584380033 podStartE2EDuration="5.029739069s" podCreationTimestamp="2025-11-24 08:03:11 +0000 UTC" firstStartedPulling="2025-11-24 08:03:12.936693281 +0000 UTC m=+4538.592675755" lastFinishedPulling="2025-11-24 08:03:15.382052317 +0000 UTC m=+4541.038034791" observedRunningTime="2025-11-24 08:03:16.027486114 +0000 UTC m=+4541.683468598" watchObservedRunningTime="2025-11-24 08:03:16.029739069 +0000 UTC m=+4541.685721563" Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.288105 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.406573 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-config\") pod \"1905c846-150f-488c-9f99-36d2f4bb985d\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.406634 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-ovsdbserver-nb\") pod \"1905c846-150f-488c-9f99-36d2f4bb985d\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.406709 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-dns-svc\") pod \"1905c846-150f-488c-9f99-36d2f4bb985d\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.406741 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxtbc\" (UniqueName: \"kubernetes.io/projected/1905c846-150f-488c-9f99-36d2f4bb985d-kube-api-access-wxtbc\") pod \"1905c846-150f-488c-9f99-36d2f4bb985d\" (UID: \"1905c846-150f-488c-9f99-36d2f4bb985d\") " Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.414909 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1905c846-150f-488c-9f99-36d2f4bb985d-kube-api-access-wxtbc" (OuterVolumeSpecName: "kube-api-access-wxtbc") pod "1905c846-150f-488c-9f99-36d2f4bb985d" (UID: "1905c846-150f-488c-9f99-36d2f4bb985d"). InnerVolumeSpecName "kube-api-access-wxtbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.429271 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-config" (OuterVolumeSpecName: "config") pod "1905c846-150f-488c-9f99-36d2f4bb985d" (UID: "1905c846-150f-488c-9f99-36d2f4bb985d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.431324 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1905c846-150f-488c-9f99-36d2f4bb985d" (UID: "1905c846-150f-488c-9f99-36d2f4bb985d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.435654 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1905c846-150f-488c-9f99-36d2f4bb985d" (UID: "1905c846-150f-488c-9f99-36d2f4bb985d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.508806 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.508857 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxtbc\" (UniqueName: \"kubernetes.io/projected/1905c846-150f-488c-9f99-36d2f4bb985d-kube-api-access-wxtbc\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.508868 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.508878 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1905c846-150f-488c-9f99-36d2f4bb985d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.993190 4799 generic.go:334] "Generic (PLEG): container finished" podID="80b671da-67db-4381-b646-d394eb8b43a0" containerID="b8a0215dfac94ab2b94f7359424eae5ed1cd437cd97d0a3f79138670bd6bdce1" exitCode=0 Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.994188 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" event={"ID":"80b671da-67db-4381-b646-d394eb8b43a0","Type":"ContainerDied","Data":"b8a0215dfac94ab2b94f7359424eae5ed1cd437cd97d0a3f79138670bd6bdce1"} Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.996752 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.996759 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b9ddc79c-884r9" event={"ID":"1905c846-150f-488c-9f99-36d2f4bb985d","Type":"ContainerDied","Data":"a8c2ca0b6252aa21a69d3dbba358dc369c9f3a0230a23f1fda745cef3ae54974"} Nov 24 08:03:16 crc kubenswrapper[4799]: I1124 08:03:16.996832 4799 scope.go:117] "RemoveContainer" containerID="5f97ff01fe1764842f25b07bfd72bc52e73c63a23a565f154797fcd3b0e6cb6b" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.200558 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57b9ddc79c-884r9"] Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.205837 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57b9ddc79c-884r9"] Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.221442 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Nov 24 08:03:17 crc kubenswrapper[4799]: E1124 08:03:17.221925 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1905c846-150f-488c-9f99-36d2f4bb985d" containerName="init" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.221941 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1905c846-150f-488c-9f99-36d2f4bb985d" containerName="init" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.222126 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1905c846-150f-488c-9f99-36d2f4bb985d" containerName="init" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.222729 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.225955 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.236193 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.325303 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f82fc\" (UniqueName: \"kubernetes.io/projected/f7836cca-6a0e-4646-9777-8d706c6cab67-kube-api-access-f82fc\") pod \"ovn-copy-data\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.325352 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/f7836cca-6a0e-4646-9777-8d706c6cab67-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.325475 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\") pod \"ovn-copy-data\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.428218 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f82fc\" (UniqueName: \"kubernetes.io/projected/f7836cca-6a0e-4646-9777-8d706c6cab67-kube-api-access-f82fc\") pod \"ovn-copy-data\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.428293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/f7836cca-6a0e-4646-9777-8d706c6cab67-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.428440 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\") pod \"ovn-copy-data\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.433359 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.433407 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\") pod \"ovn-copy-data\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e9b63d765b0fe81166c0f7d71492567b75c809832e9548ba2efa42b52e253476/globalmount\"" pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.440104 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/f7836cca-6a0e-4646-9777-8d706c6cab67-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.460976 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f82fc\" (UniqueName: \"kubernetes.io/projected/f7836cca-6a0e-4646-9777-8d706c6cab67-kube-api-access-f82fc\") pod \"ovn-copy-data\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.484867 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\") pod \"ovn-copy-data\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.548861 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 08:03:17 crc kubenswrapper[4799]: I1124 08:03:17.640324 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1905c846-150f-488c-9f99-36d2f4bb985d" path="/var/lib/kubelet/pods/1905c846-150f-488c-9f99-36d2f4bb985d/volumes" Nov 24 08:03:18 crc kubenswrapper[4799]: I1124 08:03:18.005186 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" event={"ID":"80b671da-67db-4381-b646-d394eb8b43a0","Type":"ContainerStarted","Data":"45b0d23415347550f4304ed05822ddf1584b906556c54cae8e16b61226d19444"} Nov 24 08:03:18 crc kubenswrapper[4799]: I1124 08:03:18.005338 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:18 crc kubenswrapper[4799]: I1124 08:03:18.023825 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" podStartSLOduration=4.023802153 podStartE2EDuration="4.023802153s" podCreationTimestamp="2025-11-24 08:03:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:03:18.019526941 +0000 UTC m=+4543.675509415" watchObservedRunningTime="2025-11-24 08:03:18.023802153 +0000 UTC m=+4543.679784627" Nov 24 08:03:18 crc kubenswrapper[4799]: I1124 08:03:18.124463 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 08:03:18 crc kubenswrapper[4799]: W1124 08:03:18.139370 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7836cca_6a0e_4646_9777_8d706c6cab67.slice/crio-8d4b7d62490c3a889a2fcb3d65954e7d01c3e1439d1c7ddd608dd4a6a51ea6d6 WatchSource:0}: Error finding container 8d4b7d62490c3a889a2fcb3d65954e7d01c3e1439d1c7ddd608dd4a6a51ea6d6: Status 404 returned error can't find the container with id 8d4b7d62490c3a889a2fcb3d65954e7d01c3e1439d1c7ddd608dd4a6a51ea6d6 Nov 24 08:03:19 crc kubenswrapper[4799]: I1124 08:03:19.019199 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"f7836cca-6a0e-4646-9777-8d706c6cab67","Type":"ContainerStarted","Data":"19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847"} Nov 24 08:03:19 crc kubenswrapper[4799]: I1124 08:03:19.019731 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"f7836cca-6a0e-4646-9777-8d706c6cab67","Type":"ContainerStarted","Data":"8d4b7d62490c3a889a2fcb3d65954e7d01c3e1439d1c7ddd608dd4a6a51ea6d6"} Nov 24 08:03:19 crc kubenswrapper[4799]: I1124 08:03:19.039695 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=2.8387313450000002 podStartE2EDuration="3.039663624s" podCreationTimestamp="2025-11-24 08:03:16 +0000 UTC" firstStartedPulling="2025-11-24 08:03:18.142049542 +0000 UTC m=+4543.798032006" lastFinishedPulling="2025-11-24 08:03:18.342981771 +0000 UTC m=+4543.998964285" observedRunningTime="2025-11-24 08:03:19.033727006 +0000 UTC m=+4544.689709550" watchObservedRunningTime="2025-11-24 08:03:19.039663624 +0000 UTC m=+4544.695646128" Nov 24 08:03:22 crc kubenswrapper[4799]: I1124 08:03:22.225015 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:22 crc kubenswrapper[4799]: I1124 08:03:22.225397 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:22 crc kubenswrapper[4799]: I1124 08:03:22.286691 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:23 crc kubenswrapper[4799]: I1124 08:03:23.131491 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:23 crc kubenswrapper[4799]: I1124 08:03:23.187694 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5n4k"] Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.305048 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.307037 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.310323 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.310576 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.310901 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-x8ktt" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.321731 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.446997 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7351300-d244-4acd-a82d-a2696a724d92-config\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.447049 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7351300-d244-4acd-a82d-a2696a724d92-scripts\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.447165 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnvrb\" (UniqueName: \"kubernetes.io/projected/e7351300-d244-4acd-a82d-a2696a724d92-kube-api-access-tnvrb\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.447314 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7351300-d244-4acd-a82d-a2696a724d92-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.447578 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e7351300-d244-4acd-a82d-a2696a724d92-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.549014 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7351300-d244-4acd-a82d-a2696a724d92-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.549116 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e7351300-d244-4acd-a82d-a2696a724d92-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.549139 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7351300-d244-4acd-a82d-a2696a724d92-config\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.549154 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7351300-d244-4acd-a82d-a2696a724d92-scripts\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.549219 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnvrb\" (UniqueName: \"kubernetes.io/projected/e7351300-d244-4acd-a82d-a2696a724d92-kube-api-access-tnvrb\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.551070 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7351300-d244-4acd-a82d-a2696a724d92-config\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.551067 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e7351300-d244-4acd-a82d-a2696a724d92-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.551259 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7351300-d244-4acd-a82d-a2696a724d92-scripts\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.562881 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7351300-d244-4acd-a82d-a2696a724d92-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.571722 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnvrb\" (UniqueName: \"kubernetes.io/projected/e7351300-d244-4acd-a82d-a2696a724d92-kube-api-access-tnvrb\") pod \"ovn-northd-0\" (UID: \"e7351300-d244-4acd-a82d-a2696a724d92\") " pod="openstack/ovn-northd-0" Nov 24 08:03:24 crc kubenswrapper[4799]: I1124 08:03:24.623270 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.072446 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f5n4k" podUID="1d7f6fcb-e604-4d23-b352-5de0f857f449" containerName="registry-server" containerID="cri-o://27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0" gracePeriod=2 Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.153966 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 08:03:25 crc kubenswrapper[4799]: W1124 08:03:25.171314 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7351300_d244_4acd_a82d_a2696a724d92.slice/crio-6766967d2222778a56a5709458ddd3c5459fde21d823750ccb7b21ed28c22ddf WatchSource:0}: Error finding container 6766967d2222778a56a5709458ddd3c5459fde21d823750ccb7b21ed28c22ddf: Status 404 returned error can't find the container with id 6766967d2222778a56a5709458ddd3c5459fde21d823750ccb7b21ed28c22ddf Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.310382 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.367154 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b997d949-97n87"] Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.367482 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b997d949-97n87" podUID="26016233-fb28-44e0-8490-6af2dc9bcb62" containerName="dnsmasq-dns" containerID="cri-o://7b7632164ae632dfc4974107fb150033d94f21134ecfe906e4f5bdd4ad38af0a" gracePeriod=10 Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.520077 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.672818 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-catalog-content\") pod \"1d7f6fcb-e604-4d23-b352-5de0f857f449\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.672950 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-utilities\") pod \"1d7f6fcb-e604-4d23-b352-5de0f857f449\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.673038 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdzt7\" (UniqueName: \"kubernetes.io/projected/1d7f6fcb-e604-4d23-b352-5de0f857f449-kube-api-access-kdzt7\") pod \"1d7f6fcb-e604-4d23-b352-5de0f857f449\" (UID: \"1d7f6fcb-e604-4d23-b352-5de0f857f449\") " Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.674733 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-utilities" (OuterVolumeSpecName: "utilities") pod "1d7f6fcb-e604-4d23-b352-5de0f857f449" (UID: "1d7f6fcb-e604-4d23-b352-5de0f857f449"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.692075 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d7f6fcb-e604-4d23-b352-5de0f857f449-kube-api-access-kdzt7" (OuterVolumeSpecName: "kube-api-access-kdzt7") pod "1d7f6fcb-e604-4d23-b352-5de0f857f449" (UID: "1d7f6fcb-e604-4d23-b352-5de0f857f449"). InnerVolumeSpecName "kube-api-access-kdzt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.711118 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d7f6fcb-e604-4d23-b352-5de0f857f449" (UID: "1d7f6fcb-e604-4d23-b352-5de0f857f449"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.774967 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdzt7\" (UniqueName: \"kubernetes.io/projected/1d7f6fcb-e604-4d23-b352-5de0f857f449-kube-api-access-kdzt7\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.775003 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:25 crc kubenswrapper[4799]: I1124 08:03:25.775015 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d7f6fcb-e604-4d23-b352-5de0f857f449-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.092633 4799 generic.go:334] "Generic (PLEG): container finished" podID="26016233-fb28-44e0-8490-6af2dc9bcb62" containerID="7b7632164ae632dfc4974107fb150033d94f21134ecfe906e4f5bdd4ad38af0a" exitCode=0 Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.092691 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b997d949-97n87" event={"ID":"26016233-fb28-44e0-8490-6af2dc9bcb62","Type":"ContainerDied","Data":"7b7632164ae632dfc4974107fb150033d94f21134ecfe906e4f5bdd4ad38af0a"} Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.096561 4799 generic.go:334] "Generic (PLEG): container finished" podID="1d7f6fcb-e604-4d23-b352-5de0f857f449" containerID="27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0" exitCode=0 Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.096634 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5n4k" event={"ID":"1d7f6fcb-e604-4d23-b352-5de0f857f449","Type":"ContainerDied","Data":"27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0"} Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.096667 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5n4k" event={"ID":"1d7f6fcb-e604-4d23-b352-5de0f857f449","Type":"ContainerDied","Data":"a25902e38d3763686abb16929a03d91fc82790089b458b08f818e62cd2a7a998"} Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.096683 4799 scope.go:117] "RemoveContainer" containerID="27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.096792 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5n4k" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.099352 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e7351300-d244-4acd-a82d-a2696a724d92","Type":"ContainerStarted","Data":"6766967d2222778a56a5709458ddd3c5459fde21d823750ccb7b21ed28c22ddf"} Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.238975 4799 scope.go:117] "RemoveContainer" containerID="77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.240254 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.260032 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5n4k"] Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.267331 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5n4k"] Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.281139 4799 scope.go:117] "RemoveContainer" containerID="9d595b05b9e344365250989ce0d65ae815390f7e79aaa708520df489457fe089" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.311295 4799 scope.go:117] "RemoveContainer" containerID="27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0" Nov 24 08:03:26 crc kubenswrapper[4799]: E1124 08:03:26.315051 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0\": container with ID starting with 27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0 not found: ID does not exist" containerID="27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.315097 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0"} err="failed to get container status \"27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0\": rpc error: code = NotFound desc = could not find container \"27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0\": container with ID starting with 27fe877e16ab2e0a655628bba5b9cb0425411fe0b5cc374221fdcfe5ffe824f0 not found: ID does not exist" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.315123 4799 scope.go:117] "RemoveContainer" containerID="77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910" Nov 24 08:03:26 crc kubenswrapper[4799]: E1124 08:03:26.316726 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910\": container with ID starting with 77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910 not found: ID does not exist" containerID="77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.316771 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910"} err="failed to get container status \"77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910\": rpc error: code = NotFound desc = could not find container \"77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910\": container with ID starting with 77c26bfa6a5d5305e6b83742db9e9835262f394914d35fedd175621416fc4910 not found: ID does not exist" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.316796 4799 scope.go:117] "RemoveContainer" containerID="9d595b05b9e344365250989ce0d65ae815390f7e79aaa708520df489457fe089" Nov 24 08:03:26 crc kubenswrapper[4799]: E1124 08:03:26.317224 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d595b05b9e344365250989ce0d65ae815390f7e79aaa708520df489457fe089\": container with ID starting with 9d595b05b9e344365250989ce0d65ae815390f7e79aaa708520df489457fe089 not found: ID does not exist" containerID="9d595b05b9e344365250989ce0d65ae815390f7e79aaa708520df489457fe089" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.317253 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d595b05b9e344365250989ce0d65ae815390f7e79aaa708520df489457fe089"} err="failed to get container status \"9d595b05b9e344365250989ce0d65ae815390f7e79aaa708520df489457fe089\": rpc error: code = NotFound desc = could not find container \"9d595b05b9e344365250989ce0d65ae815390f7e79aaa708520df489457fe089\": container with ID starting with 9d595b05b9e344365250989ce0d65ae815390f7e79aaa708520df489457fe089 not found: ID does not exist" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.392237 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-config\") pod \"26016233-fb28-44e0-8490-6af2dc9bcb62\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.392717 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv2dd\" (UniqueName: \"kubernetes.io/projected/26016233-fb28-44e0-8490-6af2dc9bcb62-kube-api-access-bv2dd\") pod \"26016233-fb28-44e0-8490-6af2dc9bcb62\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.392751 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-dns-svc\") pod \"26016233-fb28-44e0-8490-6af2dc9bcb62\" (UID: \"26016233-fb28-44e0-8490-6af2dc9bcb62\") " Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.396367 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26016233-fb28-44e0-8490-6af2dc9bcb62-kube-api-access-bv2dd" (OuterVolumeSpecName: "kube-api-access-bv2dd") pod "26016233-fb28-44e0-8490-6af2dc9bcb62" (UID: "26016233-fb28-44e0-8490-6af2dc9bcb62"). InnerVolumeSpecName "kube-api-access-bv2dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.429426 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26016233-fb28-44e0-8490-6af2dc9bcb62" (UID: "26016233-fb28-44e0-8490-6af2dc9bcb62"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.431407 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-config" (OuterVolumeSpecName: "config") pod "26016233-fb28-44e0-8490-6af2dc9bcb62" (UID: "26016233-fb28-44e0-8490-6af2dc9bcb62"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.494016 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv2dd\" (UniqueName: \"kubernetes.io/projected/26016233-fb28-44e0-8490-6af2dc9bcb62-kube-api-access-bv2dd\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.494174 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:26 crc kubenswrapper[4799]: I1124 08:03:26.494247 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26016233-fb28-44e0-8490-6af2dc9bcb62-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.106928 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e7351300-d244-4acd-a82d-a2696a724d92","Type":"ContainerStarted","Data":"351d4d90138f12f71c8ffe079be63ccb723b4206bc1191872dd7b98d9479e354"} Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.107271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e7351300-d244-4acd-a82d-a2696a724d92","Type":"ContainerStarted","Data":"3780d4f496e26056ae57789c7b70ccc27c184330e1014f75a92353f5927558be"} Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.107679 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.109966 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b997d949-97n87" Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.109988 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b997d949-97n87" event={"ID":"26016233-fb28-44e0-8490-6af2dc9bcb62","Type":"ContainerDied","Data":"8b341fbf9c10fc4718986fe8472de09618f4c94012ca28065b390708282d1e2e"} Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.110077 4799 scope.go:117] "RemoveContainer" containerID="7b7632164ae632dfc4974107fb150033d94f21134ecfe906e4f5bdd4ad38af0a" Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.128037 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.152808099 podStartE2EDuration="3.128015225s" podCreationTimestamp="2025-11-24 08:03:24 +0000 UTC" firstStartedPulling="2025-11-24 08:03:25.174140354 +0000 UTC m=+4550.830122828" lastFinishedPulling="2025-11-24 08:03:26.14934748 +0000 UTC m=+4551.805329954" observedRunningTime="2025-11-24 08:03:27.124461234 +0000 UTC m=+4552.780443708" watchObservedRunningTime="2025-11-24 08:03:27.128015225 +0000 UTC m=+4552.783997699" Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.136880 4799 scope.go:117] "RemoveContainer" containerID="33e59c740ed24dfff89969e0dff5aa08aae7d761799ec19347e190f722c68f60" Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.147984 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b997d949-97n87"] Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.154423 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b997d949-97n87"] Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.642877 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d7f6fcb-e604-4d23-b352-5de0f857f449" path="/var/lib/kubelet/pods/1d7f6fcb-e604-4d23-b352-5de0f857f449/volumes" Nov 24 08:03:27 crc kubenswrapper[4799]: I1124 08:03:27.644681 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26016233-fb28-44e0-8490-6af2dc9bcb62" path="/var/lib/kubelet/pods/26016233-fb28-44e0-8490-6af2dc9bcb62/volumes" Nov 24 08:03:28 crc kubenswrapper[4799]: I1124 08:03:28.640253 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:03:28 crc kubenswrapper[4799]: E1124 08:03:28.643813 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.372235 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-4d9lb"] Nov 24 08:03:29 crc kubenswrapper[4799]: E1124 08:03:29.372582 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26016233-fb28-44e0-8490-6af2dc9bcb62" containerName="init" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.372597 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="26016233-fb28-44e0-8490-6af2dc9bcb62" containerName="init" Nov 24 08:03:29 crc kubenswrapper[4799]: E1124 08:03:29.372610 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26016233-fb28-44e0-8490-6af2dc9bcb62" containerName="dnsmasq-dns" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.372616 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="26016233-fb28-44e0-8490-6af2dc9bcb62" containerName="dnsmasq-dns" Nov 24 08:03:29 crc kubenswrapper[4799]: E1124 08:03:29.372631 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7f6fcb-e604-4d23-b352-5de0f857f449" containerName="extract-content" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.372638 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7f6fcb-e604-4d23-b352-5de0f857f449" containerName="extract-content" Nov 24 08:03:29 crc kubenswrapper[4799]: E1124 08:03:29.372657 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7f6fcb-e604-4d23-b352-5de0f857f449" containerName="registry-server" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.372662 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7f6fcb-e604-4d23-b352-5de0f857f449" containerName="registry-server" Nov 24 08:03:29 crc kubenswrapper[4799]: E1124 08:03:29.372687 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7f6fcb-e604-4d23-b352-5de0f857f449" containerName="extract-utilities" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.372694 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7f6fcb-e604-4d23-b352-5de0f857f449" containerName="extract-utilities" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.372831 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d7f6fcb-e604-4d23-b352-5de0f857f449" containerName="registry-server" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.372868 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="26016233-fb28-44e0-8490-6af2dc9bcb62" containerName="dnsmasq-dns" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.373498 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4d9lb" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.395681 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-4d9lb"] Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.446824 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-operator-scripts\") pod \"keystone-db-create-4d9lb\" (UID: \"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3\") " pod="openstack/keystone-db-create-4d9lb" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.447242 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn5xh\" (UniqueName: \"kubernetes.io/projected/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-kube-api-access-xn5xh\") pod \"keystone-db-create-4d9lb\" (UID: \"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3\") " pod="openstack/keystone-db-create-4d9lb" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.488783 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-4241-account-create-vcqfd"] Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.489739 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4241-account-create-vcqfd" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.495195 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.504152 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4241-account-create-vcqfd"] Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.548745 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-operator-scripts\") pod \"keystone-db-create-4d9lb\" (UID: \"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3\") " pod="openstack/keystone-db-create-4d9lb" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.548812 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn5xh\" (UniqueName: \"kubernetes.io/projected/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-kube-api-access-xn5xh\") pod \"keystone-db-create-4d9lb\" (UID: \"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3\") " pod="openstack/keystone-db-create-4d9lb" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.549717 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-operator-scripts\") pod \"keystone-db-create-4d9lb\" (UID: \"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3\") " pod="openstack/keystone-db-create-4d9lb" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.572552 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn5xh\" (UniqueName: \"kubernetes.io/projected/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-kube-api-access-xn5xh\") pod \"keystone-db-create-4d9lb\" (UID: \"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3\") " pod="openstack/keystone-db-create-4d9lb" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.649812 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d55bs\" (UniqueName: \"kubernetes.io/projected/bbbb348f-6084-4a19-81a6-f3e5027df7fb-kube-api-access-d55bs\") pod \"keystone-4241-account-create-vcqfd\" (UID: \"bbbb348f-6084-4a19-81a6-f3e5027df7fb\") " pod="openstack/keystone-4241-account-create-vcqfd" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.649946 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbbb348f-6084-4a19-81a6-f3e5027df7fb-operator-scripts\") pod \"keystone-4241-account-create-vcqfd\" (UID: \"bbbb348f-6084-4a19-81a6-f3e5027df7fb\") " pod="openstack/keystone-4241-account-create-vcqfd" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.691685 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4d9lb" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.751751 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d55bs\" (UniqueName: \"kubernetes.io/projected/bbbb348f-6084-4a19-81a6-f3e5027df7fb-kube-api-access-d55bs\") pod \"keystone-4241-account-create-vcqfd\" (UID: \"bbbb348f-6084-4a19-81a6-f3e5027df7fb\") " pod="openstack/keystone-4241-account-create-vcqfd" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.751868 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbbb348f-6084-4a19-81a6-f3e5027df7fb-operator-scripts\") pod \"keystone-4241-account-create-vcqfd\" (UID: \"bbbb348f-6084-4a19-81a6-f3e5027df7fb\") " pod="openstack/keystone-4241-account-create-vcqfd" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.753168 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbbb348f-6084-4a19-81a6-f3e5027df7fb-operator-scripts\") pod \"keystone-4241-account-create-vcqfd\" (UID: \"bbbb348f-6084-4a19-81a6-f3e5027df7fb\") " pod="openstack/keystone-4241-account-create-vcqfd" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.770239 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d55bs\" (UniqueName: \"kubernetes.io/projected/bbbb348f-6084-4a19-81a6-f3e5027df7fb-kube-api-access-d55bs\") pod \"keystone-4241-account-create-vcqfd\" (UID: \"bbbb348f-6084-4a19-81a6-f3e5027df7fb\") " pod="openstack/keystone-4241-account-create-vcqfd" Nov 24 08:03:29 crc kubenswrapper[4799]: I1124 08:03:29.809372 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4241-account-create-vcqfd" Nov 24 08:03:30 crc kubenswrapper[4799]: I1124 08:03:30.145618 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-4d9lb"] Nov 24 08:03:30 crc kubenswrapper[4799]: W1124 08:03:30.276687 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5c93b3f_1bca_4c1a_90ff_5e8bc896dcf3.slice/crio-6f0b109bf4fa4ed7d6b18c3f6d1b40b5d30345c7d1abbdca818e5a45c8b159b6 WatchSource:0}: Error finding container 6f0b109bf4fa4ed7d6b18c3f6d1b40b5d30345c7d1abbdca818e5a45c8b159b6: Status 404 returned error can't find the container with id 6f0b109bf4fa4ed7d6b18c3f6d1b40b5d30345c7d1abbdca818e5a45c8b159b6 Nov 24 08:03:30 crc kubenswrapper[4799]: I1124 08:03:30.279557 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4241-account-create-vcqfd"] Nov 24 08:03:30 crc kubenswrapper[4799]: W1124 08:03:30.286184 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbbb348f_6084_4a19_81a6_f3e5027df7fb.slice/crio-5482ff1238a7fda442756d0c52db4d48221244da96b4700b707f64fbdbe4b88c WatchSource:0}: Error finding container 5482ff1238a7fda442756d0c52db4d48221244da96b4700b707f64fbdbe4b88c: Status 404 returned error can't find the container with id 5482ff1238a7fda442756d0c52db4d48221244da96b4700b707f64fbdbe4b88c Nov 24 08:03:31 crc kubenswrapper[4799]: I1124 08:03:31.153579 4799 generic.go:334] "Generic (PLEG): container finished" podID="b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3" containerID="75b1b31f8be72887161c747495a6c8bbd55f3a10a928ef86b2e15c85759f47c7" exitCode=0 Nov 24 08:03:31 crc kubenswrapper[4799]: I1124 08:03:31.154000 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4d9lb" event={"ID":"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3","Type":"ContainerDied","Data":"75b1b31f8be72887161c747495a6c8bbd55f3a10a928ef86b2e15c85759f47c7"} Nov 24 08:03:31 crc kubenswrapper[4799]: I1124 08:03:31.154042 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4d9lb" event={"ID":"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3","Type":"ContainerStarted","Data":"6f0b109bf4fa4ed7d6b18c3f6d1b40b5d30345c7d1abbdca818e5a45c8b159b6"} Nov 24 08:03:31 crc kubenswrapper[4799]: I1124 08:03:31.162722 4799 generic.go:334] "Generic (PLEG): container finished" podID="bbbb348f-6084-4a19-81a6-f3e5027df7fb" containerID="17595c10319016d861b3590599c798d01b512c7bd2120caecf538b9b3f03311b" exitCode=0 Nov 24 08:03:31 crc kubenswrapper[4799]: I1124 08:03:31.162797 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4241-account-create-vcqfd" event={"ID":"bbbb348f-6084-4a19-81a6-f3e5027df7fb","Type":"ContainerDied","Data":"17595c10319016d861b3590599c798d01b512c7bd2120caecf538b9b3f03311b"} Nov 24 08:03:31 crc kubenswrapper[4799]: I1124 08:03:31.162880 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4241-account-create-vcqfd" event={"ID":"bbbb348f-6084-4a19-81a6-f3e5027df7fb","Type":"ContainerStarted","Data":"5482ff1238a7fda442756d0c52db4d48221244da96b4700b707f64fbdbe4b88c"} Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.572363 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4241-account-create-vcqfd" Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.577619 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4d9lb" Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.701234 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn5xh\" (UniqueName: \"kubernetes.io/projected/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-kube-api-access-xn5xh\") pod \"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3\" (UID: \"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3\") " Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.701299 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d55bs\" (UniqueName: \"kubernetes.io/projected/bbbb348f-6084-4a19-81a6-f3e5027df7fb-kube-api-access-d55bs\") pod \"bbbb348f-6084-4a19-81a6-f3e5027df7fb\" (UID: \"bbbb348f-6084-4a19-81a6-f3e5027df7fb\") " Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.701332 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbbb348f-6084-4a19-81a6-f3e5027df7fb-operator-scripts\") pod \"bbbb348f-6084-4a19-81a6-f3e5027df7fb\" (UID: \"bbbb348f-6084-4a19-81a6-f3e5027df7fb\") " Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.701448 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-operator-scripts\") pod \"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3\" (UID: \"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3\") " Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.702379 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbbb348f-6084-4a19-81a6-f3e5027df7fb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bbbb348f-6084-4a19-81a6-f3e5027df7fb" (UID: "bbbb348f-6084-4a19-81a6-f3e5027df7fb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.702550 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3" (UID: "b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.706575 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-kube-api-access-xn5xh" (OuterVolumeSpecName: "kube-api-access-xn5xh") pod "b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3" (UID: "b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3"). InnerVolumeSpecName "kube-api-access-xn5xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.706884 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbbb348f-6084-4a19-81a6-f3e5027df7fb-kube-api-access-d55bs" (OuterVolumeSpecName: "kube-api-access-d55bs") pod "bbbb348f-6084-4a19-81a6-f3e5027df7fb" (UID: "bbbb348f-6084-4a19-81a6-f3e5027df7fb"). InnerVolumeSpecName "kube-api-access-d55bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.803374 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.803419 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn5xh\" (UniqueName: \"kubernetes.io/projected/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3-kube-api-access-xn5xh\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.803440 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d55bs\" (UniqueName: \"kubernetes.io/projected/bbbb348f-6084-4a19-81a6-f3e5027df7fb-kube-api-access-d55bs\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:32 crc kubenswrapper[4799]: I1124 08:03:32.803458 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbbb348f-6084-4a19-81a6-f3e5027df7fb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:33 crc kubenswrapper[4799]: I1124 08:03:33.184096 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4d9lb" event={"ID":"b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3","Type":"ContainerDied","Data":"6f0b109bf4fa4ed7d6b18c3f6d1b40b5d30345c7d1abbdca818e5a45c8b159b6"} Nov 24 08:03:33 crc kubenswrapper[4799]: I1124 08:03:33.184162 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f0b109bf4fa4ed7d6b18c3f6d1b40b5d30345c7d1abbdca818e5a45c8b159b6" Nov 24 08:03:33 crc kubenswrapper[4799]: I1124 08:03:33.184245 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4d9lb" Nov 24 08:03:33 crc kubenswrapper[4799]: I1124 08:03:33.189915 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4241-account-create-vcqfd" event={"ID":"bbbb348f-6084-4a19-81a6-f3e5027df7fb","Type":"ContainerDied","Data":"5482ff1238a7fda442756d0c52db4d48221244da96b4700b707f64fbdbe4b88c"} Nov 24 08:03:33 crc kubenswrapper[4799]: I1124 08:03:33.189955 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4241-account-create-vcqfd" Nov 24 08:03:33 crc kubenswrapper[4799]: I1124 08:03:33.189974 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5482ff1238a7fda442756d0c52db4d48221244da96b4700b707f64fbdbe4b88c" Nov 24 08:03:34 crc kubenswrapper[4799]: I1124 08:03:34.906003 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-56drr"] Nov 24 08:03:34 crc kubenswrapper[4799]: E1124 08:03:34.906747 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbbb348f-6084-4a19-81a6-f3e5027df7fb" containerName="mariadb-account-create" Nov 24 08:03:34 crc kubenswrapper[4799]: I1124 08:03:34.906769 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbbb348f-6084-4a19-81a6-f3e5027df7fb" containerName="mariadb-account-create" Nov 24 08:03:34 crc kubenswrapper[4799]: E1124 08:03:34.906796 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3" containerName="mariadb-database-create" Nov 24 08:03:34 crc kubenswrapper[4799]: I1124 08:03:34.906806 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3" containerName="mariadb-database-create" Nov 24 08:03:34 crc kubenswrapper[4799]: I1124 08:03:34.907087 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3" containerName="mariadb-database-create" Nov 24 08:03:34 crc kubenswrapper[4799]: I1124 08:03:34.907126 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbbb348f-6084-4a19-81a6-f3e5027df7fb" containerName="mariadb-account-create" Nov 24 08:03:34 crc kubenswrapper[4799]: I1124 08:03:34.907965 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:34 crc kubenswrapper[4799]: I1124 08:03:34.912078 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 08:03:34 crc kubenswrapper[4799]: I1124 08:03:34.912383 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 08:03:34 crc kubenswrapper[4799]: I1124 08:03:34.912541 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-cdb2z" Nov 24 08:03:34 crc kubenswrapper[4799]: I1124 08:03:34.912727 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 08:03:34 crc kubenswrapper[4799]: I1124 08:03:34.921647 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-56drr"] Nov 24 08:03:35 crc kubenswrapper[4799]: I1124 08:03:35.042545 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-config-data\") pod \"keystone-db-sync-56drr\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:35 crc kubenswrapper[4799]: I1124 08:03:35.042637 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gg8g\" (UniqueName: \"kubernetes.io/projected/7ba86795-13a5-4f7b-bc59-07a23af8b91f-kube-api-access-5gg8g\") pod \"keystone-db-sync-56drr\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:35 crc kubenswrapper[4799]: I1124 08:03:35.042738 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-combined-ca-bundle\") pod \"keystone-db-sync-56drr\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:35 crc kubenswrapper[4799]: I1124 08:03:35.145385 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-combined-ca-bundle\") pod \"keystone-db-sync-56drr\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:35 crc kubenswrapper[4799]: I1124 08:03:35.145804 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-config-data\") pod \"keystone-db-sync-56drr\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:35 crc kubenswrapper[4799]: I1124 08:03:35.145909 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gg8g\" (UniqueName: \"kubernetes.io/projected/7ba86795-13a5-4f7b-bc59-07a23af8b91f-kube-api-access-5gg8g\") pod \"keystone-db-sync-56drr\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:35 crc kubenswrapper[4799]: I1124 08:03:35.153279 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-config-data\") pod \"keystone-db-sync-56drr\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:35 crc kubenswrapper[4799]: I1124 08:03:35.153919 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-combined-ca-bundle\") pod \"keystone-db-sync-56drr\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:35 crc kubenswrapper[4799]: I1124 08:03:35.164421 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gg8g\" (UniqueName: \"kubernetes.io/projected/7ba86795-13a5-4f7b-bc59-07a23af8b91f-kube-api-access-5gg8g\") pod \"keystone-db-sync-56drr\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:35 crc kubenswrapper[4799]: I1124 08:03:35.231786 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:35 crc kubenswrapper[4799]: I1124 08:03:35.683901 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-56drr"] Nov 24 08:03:35 crc kubenswrapper[4799]: W1124 08:03:35.692564 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ba86795_13a5_4f7b_bc59_07a23af8b91f.slice/crio-bccd11c87315846cb06e2b49742d7f25fa416a7cf2b68c96c7ac4a7e7d9d240b WatchSource:0}: Error finding container bccd11c87315846cb06e2b49742d7f25fa416a7cf2b68c96c7ac4a7e7d9d240b: Status 404 returned error can't find the container with id bccd11c87315846cb06e2b49742d7f25fa416a7cf2b68c96c7ac4a7e7d9d240b Nov 24 08:03:36 crc kubenswrapper[4799]: I1124 08:03:36.216988 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-56drr" event={"ID":"7ba86795-13a5-4f7b-bc59-07a23af8b91f","Type":"ContainerStarted","Data":"bccd11c87315846cb06e2b49742d7f25fa416a7cf2b68c96c7ac4a7e7d9d240b"} Nov 24 08:03:39 crc kubenswrapper[4799]: I1124 08:03:39.684336 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 08:03:41 crc kubenswrapper[4799]: I1124 08:03:41.286239 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-56drr" event={"ID":"7ba86795-13a5-4f7b-bc59-07a23af8b91f","Type":"ContainerStarted","Data":"58f943f677764a5d1dff6bc7526a87ad6ee6c30c8c762e248f2a042890f217e2"} Nov 24 08:03:41 crc kubenswrapper[4799]: I1124 08:03:41.308503 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-56drr" podStartSLOduration=2.391774174 podStartE2EDuration="7.308484154s" podCreationTimestamp="2025-11-24 08:03:34 +0000 UTC" firstStartedPulling="2025-11-24 08:03:35.694974157 +0000 UTC m=+4561.350956631" lastFinishedPulling="2025-11-24 08:03:40.611684137 +0000 UTC m=+4566.267666611" observedRunningTime="2025-11-24 08:03:41.305284143 +0000 UTC m=+4566.961266617" watchObservedRunningTime="2025-11-24 08:03:41.308484154 +0000 UTC m=+4566.964466628" Nov 24 08:03:42 crc kubenswrapper[4799]: I1124 08:03:42.628258 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:03:42 crc kubenswrapper[4799]: E1124 08:03:42.628759 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:03:43 crc kubenswrapper[4799]: I1124 08:03:43.308585 4799 generic.go:334] "Generic (PLEG): container finished" podID="7ba86795-13a5-4f7b-bc59-07a23af8b91f" containerID="58f943f677764a5d1dff6bc7526a87ad6ee6c30c8c762e248f2a042890f217e2" exitCode=0 Nov 24 08:03:43 crc kubenswrapper[4799]: I1124 08:03:43.308639 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-56drr" event={"ID":"7ba86795-13a5-4f7b-bc59-07a23af8b91f","Type":"ContainerDied","Data":"58f943f677764a5d1dff6bc7526a87ad6ee6c30c8c762e248f2a042890f217e2"} Nov 24 08:03:44 crc kubenswrapper[4799]: I1124 08:03:44.608504 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:44 crc kubenswrapper[4799]: I1124 08:03:44.751281 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gg8g\" (UniqueName: \"kubernetes.io/projected/7ba86795-13a5-4f7b-bc59-07a23af8b91f-kube-api-access-5gg8g\") pod \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " Nov 24 08:03:44 crc kubenswrapper[4799]: I1124 08:03:44.751367 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-config-data\") pod \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " Nov 24 08:03:44 crc kubenswrapper[4799]: I1124 08:03:44.751385 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-combined-ca-bundle\") pod \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\" (UID: \"7ba86795-13a5-4f7b-bc59-07a23af8b91f\") " Nov 24 08:03:44 crc kubenswrapper[4799]: I1124 08:03:44.757831 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ba86795-13a5-4f7b-bc59-07a23af8b91f-kube-api-access-5gg8g" (OuterVolumeSpecName: "kube-api-access-5gg8g") pod "7ba86795-13a5-4f7b-bc59-07a23af8b91f" (UID: "7ba86795-13a5-4f7b-bc59-07a23af8b91f"). InnerVolumeSpecName "kube-api-access-5gg8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:03:44 crc kubenswrapper[4799]: I1124 08:03:44.791908 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ba86795-13a5-4f7b-bc59-07a23af8b91f" (UID: "7ba86795-13a5-4f7b-bc59-07a23af8b91f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:44 crc kubenswrapper[4799]: I1124 08:03:44.826786 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-config-data" (OuterVolumeSpecName: "config-data") pod "7ba86795-13a5-4f7b-bc59-07a23af8b91f" (UID: "7ba86795-13a5-4f7b-bc59-07a23af8b91f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:44 crc kubenswrapper[4799]: I1124 08:03:44.852979 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gg8g\" (UniqueName: \"kubernetes.io/projected/7ba86795-13a5-4f7b-bc59-07a23af8b91f-kube-api-access-5gg8g\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:44 crc kubenswrapper[4799]: I1124 08:03:44.853005 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:44 crc kubenswrapper[4799]: I1124 08:03:44.853014 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba86795-13a5-4f7b-bc59-07a23af8b91f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.328550 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-56drr" event={"ID":"7ba86795-13a5-4f7b-bc59-07a23af8b91f","Type":"ContainerDied","Data":"bccd11c87315846cb06e2b49742d7f25fa416a7cf2b68c96c7ac4a7e7d9d240b"} Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.328591 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bccd11c87315846cb06e2b49742d7f25fa416a7cf2b68c96c7ac4a7e7d9d240b" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.328636 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-56drr" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.627648 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f4455545-ch4nv"] Nov 24 08:03:45 crc kubenswrapper[4799]: E1124 08:03:45.628410 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ba86795-13a5-4f7b-bc59-07a23af8b91f" containerName="keystone-db-sync" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.628425 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ba86795-13a5-4f7b-bc59-07a23af8b91f" containerName="keystone-db-sync" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.628600 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ba86795-13a5-4f7b-bc59-07a23af8b91f" containerName="keystone-db-sync" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.629683 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.669286 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-s7s95"] Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.671596 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f4455545-ch4nv"] Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.672049 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s7s95"] Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.671728 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.675720 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.675820 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-cdb2z" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.677403 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.677419 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.679475 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.768274 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-dns-svc\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.768319 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-nb\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.768393 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-fernet-keys\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.768412 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-config\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.768427 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnrrp\" (UniqueName: \"kubernetes.io/projected/482dfc9d-977f-47a3-85b0-4722749aa333-kube-api-access-wnrrp\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.768443 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-config-data\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.768473 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-credential-keys\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.768490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-sb\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.768505 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-scripts\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.768527 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-combined-ca-bundle\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.768547 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-976dp\" (UniqueName: \"kubernetes.io/projected/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-kube-api-access-976dp\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.869768 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-combined-ca-bundle\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.869820 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-976dp\" (UniqueName: \"kubernetes.io/projected/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-kube-api-access-976dp\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.869910 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-dns-svc\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.869932 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-nb\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.869983 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-fernet-keys\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.870006 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-config\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.870030 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnrrp\" (UniqueName: \"kubernetes.io/projected/482dfc9d-977f-47a3-85b0-4722749aa333-kube-api-access-wnrrp\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.870050 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-config-data\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.870073 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-credential-keys\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.870092 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-sb\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.870106 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-scripts\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.871771 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-config\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.872704 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-dns-svc\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.873321 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-nb\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.874335 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-sb\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.876078 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-combined-ca-bundle\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.885293 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-scripts\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.890050 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-credential-keys\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.893440 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-fernet-keys\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.895523 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-config-data\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.906397 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-976dp\" (UniqueName: \"kubernetes.io/projected/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-kube-api-access-976dp\") pod \"keystone-bootstrap-s7s95\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.911292 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnrrp\" (UniqueName: \"kubernetes.io/projected/482dfc9d-977f-47a3-85b0-4722749aa333-kube-api-access-wnrrp\") pod \"dnsmasq-dns-7f4455545-ch4nv\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.966091 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:45 crc kubenswrapper[4799]: I1124 08:03:45.991202 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:46 crc kubenswrapper[4799]: I1124 08:03:46.445215 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s7s95"] Nov 24 08:03:46 crc kubenswrapper[4799]: I1124 08:03:46.500441 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f4455545-ch4nv"] Nov 24 08:03:46 crc kubenswrapper[4799]: W1124 08:03:46.504980 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod482dfc9d_977f_47a3_85b0_4722749aa333.slice/crio-ae44798e80098b2260de93536a1b2d842bc37a51e2c0e1e445a9064dd8c2bf99 WatchSource:0}: Error finding container ae44798e80098b2260de93536a1b2d842bc37a51e2c0e1e445a9064dd8c2bf99: Status 404 returned error can't find the container with id ae44798e80098b2260de93536a1b2d842bc37a51e2c0e1e445a9064dd8c2bf99 Nov 24 08:03:47 crc kubenswrapper[4799]: I1124 08:03:47.343963 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s7s95" event={"ID":"a91342cf-9e87-4b8f-b68b-87a7ebe5a587","Type":"ContainerStarted","Data":"5171748c56af678498262fbbe2b2360d5bccdb640b52a24f43c0b0cf2d15493a"} Nov 24 08:03:47 crc kubenswrapper[4799]: I1124 08:03:47.344018 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s7s95" event={"ID":"a91342cf-9e87-4b8f-b68b-87a7ebe5a587","Type":"ContainerStarted","Data":"317e7f55dd8bfc190ef07bafb15094ea922a8f9adb5d499c6e791c7682169cce"} Nov 24 08:03:47 crc kubenswrapper[4799]: I1124 08:03:47.346917 4799 generic.go:334] "Generic (PLEG): container finished" podID="482dfc9d-977f-47a3-85b0-4722749aa333" containerID="098fdf7b84a004818b6be1a96bf12d99ba29aa87f9a5ed15d708ce9751730c69" exitCode=0 Nov 24 08:03:47 crc kubenswrapper[4799]: I1124 08:03:47.346949 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" event={"ID":"482dfc9d-977f-47a3-85b0-4722749aa333","Type":"ContainerDied","Data":"098fdf7b84a004818b6be1a96bf12d99ba29aa87f9a5ed15d708ce9751730c69"} Nov 24 08:03:47 crc kubenswrapper[4799]: I1124 08:03:47.346987 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" event={"ID":"482dfc9d-977f-47a3-85b0-4722749aa333","Type":"ContainerStarted","Data":"ae44798e80098b2260de93536a1b2d842bc37a51e2c0e1e445a9064dd8c2bf99"} Nov 24 08:03:47 crc kubenswrapper[4799]: I1124 08:03:47.364919 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-s7s95" podStartSLOduration=2.364901625 podStartE2EDuration="2.364901625s" podCreationTimestamp="2025-11-24 08:03:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:03:47.363998159 +0000 UTC m=+4573.019980633" watchObservedRunningTime="2025-11-24 08:03:47.364901625 +0000 UTC m=+4573.020884099" Nov 24 08:03:48 crc kubenswrapper[4799]: I1124 08:03:48.358783 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" event={"ID":"482dfc9d-977f-47a3-85b0-4722749aa333","Type":"ContainerStarted","Data":"fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7"} Nov 24 08:03:49 crc kubenswrapper[4799]: I1124 08:03:49.368517 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:51 crc kubenswrapper[4799]: I1124 08:03:51.393925 4799 generic.go:334] "Generic (PLEG): container finished" podID="a91342cf-9e87-4b8f-b68b-87a7ebe5a587" containerID="5171748c56af678498262fbbe2b2360d5bccdb640b52a24f43c0b0cf2d15493a" exitCode=0 Nov 24 08:03:51 crc kubenswrapper[4799]: I1124 08:03:51.394043 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s7s95" event={"ID":"a91342cf-9e87-4b8f-b68b-87a7ebe5a587","Type":"ContainerDied","Data":"5171748c56af678498262fbbe2b2360d5bccdb640b52a24f43c0b0cf2d15493a"} Nov 24 08:03:51 crc kubenswrapper[4799]: I1124 08:03:51.418596 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" podStartSLOduration=6.418578256 podStartE2EDuration="6.418578256s" podCreationTimestamp="2025-11-24 08:03:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:03:48.385384279 +0000 UTC m=+4574.041366753" watchObservedRunningTime="2025-11-24 08:03:51.418578256 +0000 UTC m=+4577.074560730" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.733576 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.884612 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-credential-keys\") pod \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.884663 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-config-data\") pod \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.884736 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-fernet-keys\") pod \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.884767 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-combined-ca-bundle\") pod \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.884838 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-976dp\" (UniqueName: \"kubernetes.io/projected/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-kube-api-access-976dp\") pod \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.884891 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-scripts\") pod \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\" (UID: \"a91342cf-9e87-4b8f-b68b-87a7ebe5a587\") " Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.890326 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a91342cf-9e87-4b8f-b68b-87a7ebe5a587" (UID: "a91342cf-9e87-4b8f-b68b-87a7ebe5a587"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.890358 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-kube-api-access-976dp" (OuterVolumeSpecName: "kube-api-access-976dp") pod "a91342cf-9e87-4b8f-b68b-87a7ebe5a587" (UID: "a91342cf-9e87-4b8f-b68b-87a7ebe5a587"). InnerVolumeSpecName "kube-api-access-976dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.890647 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a91342cf-9e87-4b8f-b68b-87a7ebe5a587" (UID: "a91342cf-9e87-4b8f-b68b-87a7ebe5a587"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.897818 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-scripts" (OuterVolumeSpecName: "scripts") pod "a91342cf-9e87-4b8f-b68b-87a7ebe5a587" (UID: "a91342cf-9e87-4b8f-b68b-87a7ebe5a587"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.909650 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a91342cf-9e87-4b8f-b68b-87a7ebe5a587" (UID: "a91342cf-9e87-4b8f-b68b-87a7ebe5a587"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.914378 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-config-data" (OuterVolumeSpecName: "config-data") pod "a91342cf-9e87-4b8f-b68b-87a7ebe5a587" (UID: "a91342cf-9e87-4b8f-b68b-87a7ebe5a587"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.987110 4799 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.987153 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.987164 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.987176 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.987189 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-976dp\" (UniqueName: \"kubernetes.io/projected/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-kube-api-access-976dp\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:52 crc kubenswrapper[4799]: I1124 08:03:52.987202 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91342cf-9e87-4b8f-b68b-87a7ebe5a587-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.414980 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s7s95" event={"ID":"a91342cf-9e87-4b8f-b68b-87a7ebe5a587","Type":"ContainerDied","Data":"317e7f55dd8bfc190ef07bafb15094ea922a8f9adb5d499c6e791c7682169cce"} Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.415274 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="317e7f55dd8bfc190ef07bafb15094ea922a8f9adb5d499c6e791c7682169cce" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.415063 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s7s95" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.498546 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-s7s95"] Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.505055 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-s7s95"] Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.591491 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-p56f8"] Nov 24 08:03:53 crc kubenswrapper[4799]: E1124 08:03:53.591940 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a91342cf-9e87-4b8f-b68b-87a7ebe5a587" containerName="keystone-bootstrap" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.591959 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a91342cf-9e87-4b8f-b68b-87a7ebe5a587" containerName="keystone-bootstrap" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.592170 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a91342cf-9e87-4b8f-b68b-87a7ebe5a587" containerName="keystone-bootstrap" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.592866 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.596543 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.597533 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.597820 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-cdb2z" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.599578 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.600453 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.611542 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-p56f8"] Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.652504 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a91342cf-9e87-4b8f-b68b-87a7ebe5a587" path="/var/lib/kubelet/pods/a91342cf-9e87-4b8f-b68b-87a7ebe5a587/volumes" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.698978 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-scripts\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.699071 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-credential-keys\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.699099 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-combined-ca-bundle\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.699120 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-config-data\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.699135 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjkph\" (UniqueName: \"kubernetes.io/projected/3df32d62-80d5-47ab-a81c-1ce3818eb9be-kube-api-access-kjkph\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.699170 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-fernet-keys\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.800761 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-scripts\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.800810 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-credential-keys\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.800840 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-combined-ca-bundle\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.800892 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-config-data\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.800914 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjkph\" (UniqueName: \"kubernetes.io/projected/3df32d62-80d5-47ab-a81c-1ce3818eb9be-kube-api-access-kjkph\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.800967 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-fernet-keys\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.805302 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-config-data\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.805914 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-combined-ca-bundle\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.806968 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-fernet-keys\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.808195 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-scripts\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.808875 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-credential-keys\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.820121 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjkph\" (UniqueName: \"kubernetes.io/projected/3df32d62-80d5-47ab-a81c-1ce3818eb9be-kube-api-access-kjkph\") pod \"keystone-bootstrap-p56f8\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:53 crc kubenswrapper[4799]: I1124 08:03:53.917217 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:54 crc kubenswrapper[4799]: I1124 08:03:54.376544 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-p56f8"] Nov 24 08:03:54 crc kubenswrapper[4799]: I1124 08:03:54.424629 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p56f8" event={"ID":"3df32d62-80d5-47ab-a81c-1ce3818eb9be","Type":"ContainerStarted","Data":"590fdec5a91f1d357d974c330aa203c2569199dd08424f190da5924fda73dfc7"} Nov 24 08:03:55 crc kubenswrapper[4799]: I1124 08:03:55.436403 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p56f8" event={"ID":"3df32d62-80d5-47ab-a81c-1ce3818eb9be","Type":"ContainerStarted","Data":"7b14e2ac73ede37055cbdf19b9c152c448f6cebad58d91196b5539fa13d4d253"} Nov 24 08:03:55 crc kubenswrapper[4799]: I1124 08:03:55.462222 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-p56f8" podStartSLOduration=2.46220163 podStartE2EDuration="2.46220163s" podCreationTimestamp="2025-11-24 08:03:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:03:55.459084982 +0000 UTC m=+4581.115067456" watchObservedRunningTime="2025-11-24 08:03:55.46220163 +0000 UTC m=+4581.118184124" Nov 24 08:03:55 crc kubenswrapper[4799]: I1124 08:03:55.968082 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:03:56 crc kubenswrapper[4799]: I1124 08:03:56.059446 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59bcf85c7c-w569x"] Nov 24 08:03:56 crc kubenswrapper[4799]: I1124 08:03:56.059735 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" podUID="80b671da-67db-4381-b646-d394eb8b43a0" containerName="dnsmasq-dns" containerID="cri-o://45b0d23415347550f4304ed05822ddf1584b906556c54cae8e16b61226d19444" gracePeriod=10 Nov 24 08:03:56 crc kubenswrapper[4799]: I1124 08:03:56.445938 4799 generic.go:334] "Generic (PLEG): container finished" podID="80b671da-67db-4381-b646-d394eb8b43a0" containerID="45b0d23415347550f4304ed05822ddf1584b906556c54cae8e16b61226d19444" exitCode=0 Nov 24 08:03:56 crc kubenswrapper[4799]: I1124 08:03:56.446022 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" event={"ID":"80b671da-67db-4381-b646-d394eb8b43a0","Type":"ContainerDied","Data":"45b0d23415347550f4304ed05822ddf1584b906556c54cae8e16b61226d19444"} Nov 24 08:03:56 crc kubenswrapper[4799]: I1124 08:03:56.973095 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.067869 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-nb\") pod \"80b671da-67db-4381-b646-d394eb8b43a0\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.067986 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-config\") pod \"80b671da-67db-4381-b646-d394eb8b43a0\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.068013 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-sb\") pod \"80b671da-67db-4381-b646-d394eb8b43a0\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.068058 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-dns-svc\") pod \"80b671da-67db-4381-b646-d394eb8b43a0\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.068085 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7ccj\" (UniqueName: \"kubernetes.io/projected/80b671da-67db-4381-b646-d394eb8b43a0-kube-api-access-g7ccj\") pod \"80b671da-67db-4381-b646-d394eb8b43a0\" (UID: \"80b671da-67db-4381-b646-d394eb8b43a0\") " Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.080993 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b671da-67db-4381-b646-d394eb8b43a0-kube-api-access-g7ccj" (OuterVolumeSpecName: "kube-api-access-g7ccj") pod "80b671da-67db-4381-b646-d394eb8b43a0" (UID: "80b671da-67db-4381-b646-d394eb8b43a0"). InnerVolumeSpecName "kube-api-access-g7ccj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.111539 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "80b671da-67db-4381-b646-d394eb8b43a0" (UID: "80b671da-67db-4381-b646-d394eb8b43a0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.121326 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "80b671da-67db-4381-b646-d394eb8b43a0" (UID: "80b671da-67db-4381-b646-d394eb8b43a0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.133420 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "80b671da-67db-4381-b646-d394eb8b43a0" (UID: "80b671da-67db-4381-b646-d394eb8b43a0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.141550 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-config" (OuterVolumeSpecName: "config") pod "80b671da-67db-4381-b646-d394eb8b43a0" (UID: "80b671da-67db-4381-b646-d394eb8b43a0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.170085 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.170120 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.170131 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.170140 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80b671da-67db-4381-b646-d394eb8b43a0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.170150 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7ccj\" (UniqueName: \"kubernetes.io/projected/80b671da-67db-4381-b646-d394eb8b43a0-kube-api-access-g7ccj\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.459282 4799 generic.go:334] "Generic (PLEG): container finished" podID="3df32d62-80d5-47ab-a81c-1ce3818eb9be" containerID="7b14e2ac73ede37055cbdf19b9c152c448f6cebad58d91196b5539fa13d4d253" exitCode=0 Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.459355 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p56f8" event={"ID":"3df32d62-80d5-47ab-a81c-1ce3818eb9be","Type":"ContainerDied","Data":"7b14e2ac73ede37055cbdf19b9c152c448f6cebad58d91196b5539fa13d4d253"} Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.462074 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" event={"ID":"80b671da-67db-4381-b646-d394eb8b43a0","Type":"ContainerDied","Data":"3c8353a616628d1136ed0e6207e88b9283b85258f8f53ccd6cb6aa98057efaff"} Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.462112 4799 scope.go:117] "RemoveContainer" containerID="45b0d23415347550f4304ed05822ddf1584b906556c54cae8e16b61226d19444" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.462229 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59bcf85c7c-w569x" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.507550 4799 scope.go:117] "RemoveContainer" containerID="b8a0215dfac94ab2b94f7359424eae5ed1cd437cd97d0a3f79138670bd6bdce1" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.530420 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59bcf85c7c-w569x"] Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.541076 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59bcf85c7c-w569x"] Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.628431 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:03:57 crc kubenswrapper[4799]: I1124 08:03:57.637931 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80b671da-67db-4381-b646-d394eb8b43a0" path="/var/lib/kubelet/pods/80b671da-67db-4381-b646-d394eb8b43a0/volumes" Nov 24 08:03:58 crc kubenswrapper[4799]: I1124 08:03:58.475072 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"023b79a79ebb461005d32f47310810404d347caf336f616e7c3a30dfcc69e291"} Nov 24 08:03:58 crc kubenswrapper[4799]: I1124 08:03:58.835643 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.012307 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-fernet-keys\") pod \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.012358 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-combined-ca-bundle\") pod \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.012407 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-scripts\") pod \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.012457 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-config-data\") pod \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.012490 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjkph\" (UniqueName: \"kubernetes.io/projected/3df32d62-80d5-47ab-a81c-1ce3818eb9be-kube-api-access-kjkph\") pod \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.012520 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-credential-keys\") pod \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\" (UID: \"3df32d62-80d5-47ab-a81c-1ce3818eb9be\") " Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.018130 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3df32d62-80d5-47ab-a81c-1ce3818eb9be-kube-api-access-kjkph" (OuterVolumeSpecName: "kube-api-access-kjkph") pod "3df32d62-80d5-47ab-a81c-1ce3818eb9be" (UID: "3df32d62-80d5-47ab-a81c-1ce3818eb9be"). InnerVolumeSpecName "kube-api-access-kjkph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.019156 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-scripts" (OuterVolumeSpecName: "scripts") pod "3df32d62-80d5-47ab-a81c-1ce3818eb9be" (UID: "3df32d62-80d5-47ab-a81c-1ce3818eb9be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.020022 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3df32d62-80d5-47ab-a81c-1ce3818eb9be" (UID: "3df32d62-80d5-47ab-a81c-1ce3818eb9be"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.022951 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3df32d62-80d5-47ab-a81c-1ce3818eb9be" (UID: "3df32d62-80d5-47ab-a81c-1ce3818eb9be"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.036286 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-config-data" (OuterVolumeSpecName: "config-data") pod "3df32d62-80d5-47ab-a81c-1ce3818eb9be" (UID: "3df32d62-80d5-47ab-a81c-1ce3818eb9be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.041330 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3df32d62-80d5-47ab-a81c-1ce3818eb9be" (UID: "3df32d62-80d5-47ab-a81c-1ce3818eb9be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.114057 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.114101 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjkph\" (UniqueName: \"kubernetes.io/projected/3df32d62-80d5-47ab-a81c-1ce3818eb9be-kube-api-access-kjkph\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.114122 4799 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.114138 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.114157 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.114174 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3df32d62-80d5-47ab-a81c-1ce3818eb9be-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.486304 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p56f8" event={"ID":"3df32d62-80d5-47ab-a81c-1ce3818eb9be","Type":"ContainerDied","Data":"590fdec5a91f1d357d974c330aa203c2569199dd08424f190da5924fda73dfc7"} Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.486426 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p56f8" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.490230 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="590fdec5a91f1d357d974c330aa203c2569199dd08424f190da5924fda73dfc7" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.572860 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-658ddd8bf4-qjlrl"] Nov 24 08:03:59 crc kubenswrapper[4799]: E1124 08:03:59.573453 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b671da-67db-4381-b646-d394eb8b43a0" containerName="dnsmasq-dns" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.573469 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b671da-67db-4381-b646-d394eb8b43a0" containerName="dnsmasq-dns" Nov 24 08:03:59 crc kubenswrapper[4799]: E1124 08:03:59.573495 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b671da-67db-4381-b646-d394eb8b43a0" containerName="init" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.573503 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b671da-67db-4381-b646-d394eb8b43a0" containerName="init" Nov 24 08:03:59 crc kubenswrapper[4799]: E1124 08:03:59.573515 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df32d62-80d5-47ab-a81c-1ce3818eb9be" containerName="keystone-bootstrap" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.573521 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df32d62-80d5-47ab-a81c-1ce3818eb9be" containerName="keystone-bootstrap" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.573689 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b671da-67db-4381-b646-d394eb8b43a0" containerName="dnsmasq-dns" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.573716 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df32d62-80d5-47ab-a81c-1ce3818eb9be" containerName="keystone-bootstrap" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.574259 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.578497 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.578710 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.578866 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.578995 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-cdb2z" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.582898 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-658ddd8bf4-qjlrl"] Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.723413 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-combined-ca-bundle\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.723489 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-credential-keys\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.723515 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-scripts\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.723547 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-fernet-keys\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.723579 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgcpr\" (UniqueName: \"kubernetes.io/projected/a963eddd-190c-4a8b-adda-9971855f9beb-kube-api-access-dgcpr\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.723773 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-config-data\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.825447 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-scripts\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.825522 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-fernet-keys\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.825550 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgcpr\" (UniqueName: \"kubernetes.io/projected/a963eddd-190c-4a8b-adda-9971855f9beb-kube-api-access-dgcpr\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.825592 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-config-data\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.825662 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-combined-ca-bundle\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.825694 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-credential-keys\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.830111 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-fernet-keys\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.830171 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-credential-keys\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.830287 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-combined-ca-bundle\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.833226 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-scripts\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.833513 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a963eddd-190c-4a8b-adda-9971855f9beb-config-data\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.843296 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgcpr\" (UniqueName: \"kubernetes.io/projected/a963eddd-190c-4a8b-adda-9971855f9beb-kube-api-access-dgcpr\") pod \"keystone-658ddd8bf4-qjlrl\" (UID: \"a963eddd-190c-4a8b-adda-9971855f9beb\") " pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:03:59 crc kubenswrapper[4799]: I1124 08:03:59.907202 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:04:00 crc kubenswrapper[4799]: I1124 08:04:00.355141 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-658ddd8bf4-qjlrl"] Nov 24 08:04:00 crc kubenswrapper[4799]: W1124 08:04:00.356679 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda963eddd_190c_4a8b_adda_9971855f9beb.slice/crio-ee2d6d09a00795662fe21e7a60eb2e737977864563c752947f27dc677d04f08f WatchSource:0}: Error finding container ee2d6d09a00795662fe21e7a60eb2e737977864563c752947f27dc677d04f08f: Status 404 returned error can't find the container with id ee2d6d09a00795662fe21e7a60eb2e737977864563c752947f27dc677d04f08f Nov 24 08:04:00 crc kubenswrapper[4799]: I1124 08:04:00.500583 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-658ddd8bf4-qjlrl" event={"ID":"a963eddd-190c-4a8b-adda-9971855f9beb","Type":"ContainerStarted","Data":"ee2d6d09a00795662fe21e7a60eb2e737977864563c752947f27dc677d04f08f"} Nov 24 08:04:01 crc kubenswrapper[4799]: I1124 08:04:01.509764 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-658ddd8bf4-qjlrl" event={"ID":"a963eddd-190c-4a8b-adda-9971855f9beb","Type":"ContainerStarted","Data":"8d4c01b0b07cd0fac150abe5d40f958ac7e65e4f3582c349c6ac03c5b43724d7"} Nov 24 08:04:01 crc kubenswrapper[4799]: I1124 08:04:01.510138 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:04:01 crc kubenswrapper[4799]: I1124 08:04:01.531281 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-658ddd8bf4-qjlrl" podStartSLOduration=2.531249541 podStartE2EDuration="2.531249541s" podCreationTimestamp="2025-11-24 08:03:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:04:01.528515603 +0000 UTC m=+4587.184498077" watchObservedRunningTime="2025-11-24 08:04:01.531249541 +0000 UTC m=+4587.187232045" Nov 24 08:04:31 crc kubenswrapper[4799]: I1124 08:04:31.316635 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-658ddd8bf4-qjlrl" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.513610 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.515653 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.518280 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-7sl5h" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.519469 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.519510 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.529501 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.595527 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-748sk"] Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.600033 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.623526 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-748sk"] Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.686334 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5e50e296-2811-4760-a259-86377d94888b-openstack-config\") pod \"openstackclient\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " pod="openstack/openstackclient" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.686402 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bw6m\" (UniqueName: \"kubernetes.io/projected/5e50e296-2811-4760-a259-86377d94888b-kube-api-access-8bw6m\") pod \"openstackclient\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " pod="openstack/openstackclient" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.686436 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5e50e296-2811-4760-a259-86377d94888b-openstack-config-secret\") pod \"openstackclient\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " pod="openstack/openstackclient" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.787584 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bw6m\" (UniqueName: \"kubernetes.io/projected/5e50e296-2811-4760-a259-86377d94888b-kube-api-access-8bw6m\") pod \"openstackclient\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " pod="openstack/openstackclient" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.787690 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5e50e296-2811-4760-a259-86377d94888b-openstack-config-secret\") pod \"openstackclient\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " pod="openstack/openstackclient" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.787779 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-utilities\") pod \"community-operators-748sk\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.787921 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gcnz\" (UniqueName: \"kubernetes.io/projected/1f1d263a-eabf-400d-9453-8dc44c131385-kube-api-access-2gcnz\") pod \"community-operators-748sk\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.787950 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5e50e296-2811-4760-a259-86377d94888b-openstack-config\") pod \"openstackclient\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " pod="openstack/openstackclient" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.787999 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-catalog-content\") pod \"community-operators-748sk\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.788896 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5e50e296-2811-4760-a259-86377d94888b-openstack-config\") pod \"openstackclient\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " pod="openstack/openstackclient" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.793688 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5e50e296-2811-4760-a259-86377d94888b-openstack-config-secret\") pod \"openstackclient\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " pod="openstack/openstackclient" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.810434 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bw6m\" (UniqueName: \"kubernetes.io/projected/5e50e296-2811-4760-a259-86377d94888b-kube-api-access-8bw6m\") pod \"openstackclient\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " pod="openstack/openstackclient" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.845595 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.889759 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-catalog-content\") pod \"community-operators-748sk\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.890183 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-utilities\") pod \"community-operators-748sk\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.890284 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gcnz\" (UniqueName: \"kubernetes.io/projected/1f1d263a-eabf-400d-9453-8dc44c131385-kube-api-access-2gcnz\") pod \"community-operators-748sk\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.890537 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-catalog-content\") pod \"community-operators-748sk\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.892330 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-utilities\") pod \"community-operators-748sk\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.910763 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gcnz\" (UniqueName: \"kubernetes.io/projected/1f1d263a-eabf-400d-9453-8dc44c131385-kube-api-access-2gcnz\") pod \"community-operators-748sk\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:34 crc kubenswrapper[4799]: I1124 08:04:34.926077 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:35 crc kubenswrapper[4799]: I1124 08:04:35.134099 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 08:04:35 crc kubenswrapper[4799]: I1124 08:04:35.475740 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-748sk"] Nov 24 08:04:35 crc kubenswrapper[4799]: W1124 08:04:35.483757 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-143844553ca46d632ea3d706a6c51c0ef43a856bd451d24995d7f140da5c96b3 WatchSource:0}: Error finding container 143844553ca46d632ea3d706a6c51c0ef43a856bd451d24995d7f140da5c96b3: Status 404 returned error can't find the container with id 143844553ca46d632ea3d706a6c51c0ef43a856bd451d24995d7f140da5c96b3 Nov 24 08:04:35 crc kubenswrapper[4799]: I1124 08:04:35.853160 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5e50e296-2811-4760-a259-86377d94888b","Type":"ContainerStarted","Data":"022bed948c3e937661b18a27aa78e54fe490ac3940d0fa1f427a30f8260feaaa"} Nov 24 08:04:35 crc kubenswrapper[4799]: I1124 08:04:35.856945 4799 generic.go:334] "Generic (PLEG): container finished" podID="1f1d263a-eabf-400d-9453-8dc44c131385" containerID="57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856" exitCode=0 Nov 24 08:04:35 crc kubenswrapper[4799]: I1124 08:04:35.857021 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-748sk" event={"ID":"1f1d263a-eabf-400d-9453-8dc44c131385","Type":"ContainerDied","Data":"57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856"} Nov 24 08:04:35 crc kubenswrapper[4799]: I1124 08:04:35.857479 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-748sk" event={"ID":"1f1d263a-eabf-400d-9453-8dc44c131385","Type":"ContainerStarted","Data":"143844553ca46d632ea3d706a6c51c0ef43a856bd451d24995d7f140da5c96b3"} Nov 24 08:04:37 crc kubenswrapper[4799]: I1124 08:04:37.899560 4799 generic.go:334] "Generic (PLEG): container finished" podID="1f1d263a-eabf-400d-9453-8dc44c131385" containerID="ccd328cf6ceac672d1b5b4c8ef4207aa3dade5c69f7552a843708d416fcf9169" exitCode=0 Nov 24 08:04:37 crc kubenswrapper[4799]: I1124 08:04:37.899624 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-748sk" event={"ID":"1f1d263a-eabf-400d-9453-8dc44c131385","Type":"ContainerDied","Data":"ccd328cf6ceac672d1b5b4c8ef4207aa3dade5c69f7552a843708d416fcf9169"} Nov 24 08:04:38 crc kubenswrapper[4799]: I1124 08:04:38.910560 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-748sk" event={"ID":"1f1d263a-eabf-400d-9453-8dc44c131385","Type":"ContainerStarted","Data":"c270f4aab6347084cab68168dc075cf6445a6789bfc7a63af95a8ef214d58f90"} Nov 24 08:04:43 crc kubenswrapper[4799]: E1124 08:04:43.711881 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-conmon-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:04:44 crc kubenswrapper[4799]: I1124 08:04:44.926712 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:44 crc kubenswrapper[4799]: I1124 08:04:44.927091 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:45 crc kubenswrapper[4799]: I1124 08:04:45.016999 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:45 crc kubenswrapper[4799]: I1124 08:04:45.039383 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-748sk" podStartSLOduration=8.584537013 podStartE2EDuration="11.039357697s" podCreationTimestamp="2025-11-24 08:04:34 +0000 UTC" firstStartedPulling="2025-11-24 08:04:35.860115034 +0000 UTC m=+4621.516097548" lastFinishedPulling="2025-11-24 08:04:38.314935758 +0000 UTC m=+4623.970918232" observedRunningTime="2025-11-24 08:04:38.930462476 +0000 UTC m=+4624.586444950" watchObservedRunningTime="2025-11-24 08:04:45.039357697 +0000 UTC m=+4630.695340161" Nov 24 08:04:45 crc kubenswrapper[4799]: I1124 08:04:45.062159 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:45 crc kubenswrapper[4799]: I1124 08:04:45.252647 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-748sk"] Nov 24 08:04:46 crc kubenswrapper[4799]: I1124 08:04:46.980664 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-748sk" podUID="1f1d263a-eabf-400d-9453-8dc44c131385" containerName="registry-server" containerID="cri-o://c270f4aab6347084cab68168dc075cf6445a6789bfc7a63af95a8ef214d58f90" gracePeriod=2 Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.017204 4799 generic.go:334] "Generic (PLEG): container finished" podID="1f1d263a-eabf-400d-9453-8dc44c131385" containerID="c270f4aab6347084cab68168dc075cf6445a6789bfc7a63af95a8ef214d58f90" exitCode=0 Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.017281 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-748sk" event={"ID":"1f1d263a-eabf-400d-9453-8dc44c131385","Type":"ContainerDied","Data":"c270f4aab6347084cab68168dc075cf6445a6789bfc7a63af95a8ef214d58f90"} Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.213186 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.246535 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-catalog-content\") pod \"1f1d263a-eabf-400d-9453-8dc44c131385\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.246613 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gcnz\" (UniqueName: \"kubernetes.io/projected/1f1d263a-eabf-400d-9453-8dc44c131385-kube-api-access-2gcnz\") pod \"1f1d263a-eabf-400d-9453-8dc44c131385\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.246647 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-utilities\") pod \"1f1d263a-eabf-400d-9453-8dc44c131385\" (UID: \"1f1d263a-eabf-400d-9453-8dc44c131385\") " Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.247971 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-utilities" (OuterVolumeSpecName: "utilities") pod "1f1d263a-eabf-400d-9453-8dc44c131385" (UID: "1f1d263a-eabf-400d-9453-8dc44c131385"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.251044 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f1d263a-eabf-400d-9453-8dc44c131385-kube-api-access-2gcnz" (OuterVolumeSpecName: "kube-api-access-2gcnz") pod "1f1d263a-eabf-400d-9453-8dc44c131385" (UID: "1f1d263a-eabf-400d-9453-8dc44c131385"). InnerVolumeSpecName "kube-api-access-2gcnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.300367 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f1d263a-eabf-400d-9453-8dc44c131385" (UID: "1f1d263a-eabf-400d-9453-8dc44c131385"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.348730 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gcnz\" (UniqueName: \"kubernetes.io/projected/1f1d263a-eabf-400d-9453-8dc44c131385-kube-api-access-2gcnz\") on node \"crc\" DevicePath \"\"" Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.348776 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:04:48 crc kubenswrapper[4799]: I1124 08:04:48.348790 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f1d263a-eabf-400d-9453-8dc44c131385-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:04:49 crc kubenswrapper[4799]: I1124 08:04:49.031959 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-748sk" event={"ID":"1f1d263a-eabf-400d-9453-8dc44c131385","Type":"ContainerDied","Data":"143844553ca46d632ea3d706a6c51c0ef43a856bd451d24995d7f140da5c96b3"} Nov 24 08:04:49 crc kubenswrapper[4799]: I1124 08:04:49.031987 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-748sk" Nov 24 08:04:49 crc kubenswrapper[4799]: I1124 08:04:49.032017 4799 scope.go:117] "RemoveContainer" containerID="c270f4aab6347084cab68168dc075cf6445a6789bfc7a63af95a8ef214d58f90" Nov 24 08:04:49 crc kubenswrapper[4799]: I1124 08:04:49.033640 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5e50e296-2811-4760-a259-86377d94888b","Type":"ContainerStarted","Data":"aa8be354bf5825ffaa9fa768910a05d44954e65d34f59d8f186ecf39dce49c04"} Nov 24 08:04:49 crc kubenswrapper[4799]: I1124 08:04:49.049968 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.218805863 podStartE2EDuration="15.049947324s" podCreationTimestamp="2025-11-24 08:04:34 +0000 UTC" firstStartedPulling="2025-11-24 08:04:35.154828725 +0000 UTC m=+4620.810811199" lastFinishedPulling="2025-11-24 08:04:47.985970186 +0000 UTC m=+4633.641952660" observedRunningTime="2025-11-24 08:04:49.048973496 +0000 UTC m=+4634.704955970" watchObservedRunningTime="2025-11-24 08:04:49.049947324 +0000 UTC m=+4634.705929828" Nov 24 08:04:49 crc kubenswrapper[4799]: I1124 08:04:49.054080 4799 scope.go:117] "RemoveContainer" containerID="ccd328cf6ceac672d1b5b4c8ef4207aa3dade5c69f7552a843708d416fcf9169" Nov 24 08:04:49 crc kubenswrapper[4799]: I1124 08:04:49.073004 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-748sk"] Nov 24 08:04:49 crc kubenswrapper[4799]: I1124 08:04:49.082820 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-748sk"] Nov 24 08:04:49 crc kubenswrapper[4799]: I1124 08:04:49.084952 4799 scope.go:117] "RemoveContainer" containerID="57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856" Nov 24 08:04:49 crc kubenswrapper[4799]: I1124 08:04:49.642971 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f1d263a-eabf-400d-9453-8dc44c131385" path="/var/lib/kubelet/pods/1f1d263a-eabf-400d-9453-8dc44c131385/volumes" Nov 24 08:04:53 crc kubenswrapper[4799]: E1124 08:04:53.928285 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-conmon-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:04:54 crc kubenswrapper[4799]: I1124 08:04:54.495631 4799 scope.go:117] "RemoveContainer" containerID="93629a7205f9f1bc69fa835ffa37243c0154ab48ee6e86d52d814e2dabbe445b" Nov 24 08:05:04 crc kubenswrapper[4799]: E1124 08:05:04.135765 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-conmon-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:05:14 crc kubenswrapper[4799]: E1124 08:05:14.353297 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-conmon-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:05:16 crc kubenswrapper[4799]: E1124 08:05:16.335629 4799 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:34522->38.102.83.50:45263: write tcp 38.102.83.50:34522->38.102.83.50:45263: write: broken pipe Nov 24 08:05:24 crc kubenswrapper[4799]: E1124 08:05:24.600806 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-conmon-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:05:34 crc kubenswrapper[4799]: E1124 08:05:34.802489 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f1d263a_eabf_400d_9453_8dc44c131385.slice/crio-conmon-57ac77bb1df8f627b4910cc693542c66023fc346e5adceed32747af5c9b75856.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:05:54 crc kubenswrapper[4799]: I1124 08:05:54.589587 4799 scope.go:117] "RemoveContainer" containerID="d33c246e9b2dc9a3e6c2b9c329f572a2598bbc9753f54f639bcefc377a192cb4" Nov 24 08:05:54 crc kubenswrapper[4799]: I1124 08:05:54.609407 4799 scope.go:117] "RemoveContainer" containerID="0356867f9d47ad29a15da72d5d230d72c1ded27aed899d8a9e0a67454eea719f" Nov 24 08:05:54 crc kubenswrapper[4799]: I1124 08:05:54.657940 4799 scope.go:117] "RemoveContainer" containerID="0c7e10862a39a5a52d591852b7ce843d5388fb7dbccf635e685b637b483b1a59" Nov 24 08:05:54 crc kubenswrapper[4799]: I1124 08:05:54.699994 4799 scope.go:117] "RemoveContainer" containerID="e8573b08f19e25d030ab4588a0906815dc5827afa37b711ab2339e8e2da65a9f" Nov 24 08:05:54 crc kubenswrapper[4799]: I1124 08:05:54.731832 4799 scope.go:117] "RemoveContainer" containerID="36326a5305ebb27050b9faeb368319a3091f2989685d5a9009fb90dd71518b8b" Nov 24 08:05:54 crc kubenswrapper[4799]: I1124 08:05:54.762696 4799 scope.go:117] "RemoveContainer" containerID="6de72d6906d6946f26d69d189df9b8cad79db64be0f5da3ad22cc3770925d5c9" Nov 24 08:05:54 crc kubenswrapper[4799]: I1124 08:05:54.794373 4799 scope.go:117] "RemoveContainer" containerID="8ed399e008d8244f6f977ed7d84630cfb86a2e4d4d3dfde3f3cf0049e52f6ad4" Nov 24 08:05:59 crc kubenswrapper[4799]: E1124 08:05:59.448102 4799 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:35348->38.102.83.50:45263: write tcp 38.102.83.50:35348->38.102.83.50:45263: write: broken pipe Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.178456 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-dtcc5"] Nov 24 08:06:07 crc kubenswrapper[4799]: E1124 08:06:07.179635 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1d263a-eabf-400d-9453-8dc44c131385" containerName="extract-content" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.179654 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1d263a-eabf-400d-9453-8dc44c131385" containerName="extract-content" Nov 24 08:06:07 crc kubenswrapper[4799]: E1124 08:06:07.179682 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1d263a-eabf-400d-9453-8dc44c131385" containerName="registry-server" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.179692 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1d263a-eabf-400d-9453-8dc44c131385" containerName="registry-server" Nov 24 08:06:07 crc kubenswrapper[4799]: E1124 08:06:07.179717 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1d263a-eabf-400d-9453-8dc44c131385" containerName="extract-utilities" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.179726 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1d263a-eabf-400d-9453-8dc44c131385" containerName="extract-utilities" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.179956 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f1d263a-eabf-400d-9453-8dc44c131385" containerName="registry-server" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.180705 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dtcc5" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.196168 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dtcc5"] Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.241131 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pwns\" (UniqueName: \"kubernetes.io/projected/4c5723e9-24ac-4cfa-b181-7030c18c2d27-kube-api-access-4pwns\") pod \"barbican-db-create-dtcc5\" (UID: \"4c5723e9-24ac-4cfa-b181-7030c18c2d27\") " pod="openstack/barbican-db-create-dtcc5" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.241299 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c5723e9-24ac-4cfa-b181-7030c18c2d27-operator-scripts\") pod \"barbican-db-create-dtcc5\" (UID: \"4c5723e9-24ac-4cfa-b181-7030c18c2d27\") " pod="openstack/barbican-db-create-dtcc5" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.269395 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-d14e-account-create-9lhsh"] Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.270695 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d14e-account-create-9lhsh" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.272529 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.277875 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d14e-account-create-9lhsh"] Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.342940 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c5723e9-24ac-4cfa-b181-7030c18c2d27-operator-scripts\") pod \"barbican-db-create-dtcc5\" (UID: \"4c5723e9-24ac-4cfa-b181-7030c18c2d27\") " pod="openstack/barbican-db-create-dtcc5" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.343008 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/508bd06c-03d1-44fe-9425-5a36f40a60b5-operator-scripts\") pod \"barbican-d14e-account-create-9lhsh\" (UID: \"508bd06c-03d1-44fe-9425-5a36f40a60b5\") " pod="openstack/barbican-d14e-account-create-9lhsh" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.343055 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pwns\" (UniqueName: \"kubernetes.io/projected/4c5723e9-24ac-4cfa-b181-7030c18c2d27-kube-api-access-4pwns\") pod \"barbican-db-create-dtcc5\" (UID: \"4c5723e9-24ac-4cfa-b181-7030c18c2d27\") " pod="openstack/barbican-db-create-dtcc5" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.343093 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzgrt\" (UniqueName: \"kubernetes.io/projected/508bd06c-03d1-44fe-9425-5a36f40a60b5-kube-api-access-hzgrt\") pod \"barbican-d14e-account-create-9lhsh\" (UID: \"508bd06c-03d1-44fe-9425-5a36f40a60b5\") " pod="openstack/barbican-d14e-account-create-9lhsh" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.343764 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c5723e9-24ac-4cfa-b181-7030c18c2d27-operator-scripts\") pod \"barbican-db-create-dtcc5\" (UID: \"4c5723e9-24ac-4cfa-b181-7030c18c2d27\") " pod="openstack/barbican-db-create-dtcc5" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.363930 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pwns\" (UniqueName: \"kubernetes.io/projected/4c5723e9-24ac-4cfa-b181-7030c18c2d27-kube-api-access-4pwns\") pod \"barbican-db-create-dtcc5\" (UID: \"4c5723e9-24ac-4cfa-b181-7030c18c2d27\") " pod="openstack/barbican-db-create-dtcc5" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.443794 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/508bd06c-03d1-44fe-9425-5a36f40a60b5-operator-scripts\") pod \"barbican-d14e-account-create-9lhsh\" (UID: \"508bd06c-03d1-44fe-9425-5a36f40a60b5\") " pod="openstack/barbican-d14e-account-create-9lhsh" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.444248 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzgrt\" (UniqueName: \"kubernetes.io/projected/508bd06c-03d1-44fe-9425-5a36f40a60b5-kube-api-access-hzgrt\") pod \"barbican-d14e-account-create-9lhsh\" (UID: \"508bd06c-03d1-44fe-9425-5a36f40a60b5\") " pod="openstack/barbican-d14e-account-create-9lhsh" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.445949 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/508bd06c-03d1-44fe-9425-5a36f40a60b5-operator-scripts\") pod \"barbican-d14e-account-create-9lhsh\" (UID: \"508bd06c-03d1-44fe-9425-5a36f40a60b5\") " pod="openstack/barbican-d14e-account-create-9lhsh" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.463537 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzgrt\" (UniqueName: \"kubernetes.io/projected/508bd06c-03d1-44fe-9425-5a36f40a60b5-kube-api-access-hzgrt\") pod \"barbican-d14e-account-create-9lhsh\" (UID: \"508bd06c-03d1-44fe-9425-5a36f40a60b5\") " pod="openstack/barbican-d14e-account-create-9lhsh" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.510980 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dtcc5" Nov 24 08:06:07 crc kubenswrapper[4799]: I1124 08:06:07.587595 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d14e-account-create-9lhsh" Nov 24 08:06:08 crc kubenswrapper[4799]: I1124 08:06:08.027793 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dtcc5"] Nov 24 08:06:08 crc kubenswrapper[4799]: I1124 08:06:08.071071 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d14e-account-create-9lhsh"] Nov 24 08:06:08 crc kubenswrapper[4799]: W1124 08:06:08.074884 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod508bd06c_03d1_44fe_9425_5a36f40a60b5.slice/crio-f743fd2548ec3416cbcd22b97a82fc9aff3153e7f8d9ef2ca74db2b8d08307fe WatchSource:0}: Error finding container f743fd2548ec3416cbcd22b97a82fc9aff3153e7f8d9ef2ca74db2b8d08307fe: Status 404 returned error can't find the container with id f743fd2548ec3416cbcd22b97a82fc9aff3153e7f8d9ef2ca74db2b8d08307fe Nov 24 08:06:08 crc kubenswrapper[4799]: I1124 08:06:08.794234 4799 generic.go:334] "Generic (PLEG): container finished" podID="4c5723e9-24ac-4cfa-b181-7030c18c2d27" containerID="ddb8609e7c37cf623fef37f8ed548cece157e70125cfd22887adb730d639cd6b" exitCode=0 Nov 24 08:06:08 crc kubenswrapper[4799]: I1124 08:06:08.794331 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dtcc5" event={"ID":"4c5723e9-24ac-4cfa-b181-7030c18c2d27","Type":"ContainerDied","Data":"ddb8609e7c37cf623fef37f8ed548cece157e70125cfd22887adb730d639cd6b"} Nov 24 08:06:08 crc kubenswrapper[4799]: I1124 08:06:08.794768 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dtcc5" event={"ID":"4c5723e9-24ac-4cfa-b181-7030c18c2d27","Type":"ContainerStarted","Data":"eb62f4f4fe51de9e5aafe84de3985d3df4f757f36b31fbdaf15d5ce2ff10b413"} Nov 24 08:06:08 crc kubenswrapper[4799]: I1124 08:06:08.796784 4799 generic.go:334] "Generic (PLEG): container finished" podID="508bd06c-03d1-44fe-9425-5a36f40a60b5" containerID="cb9401eac7115b5bcb4c506c146c4ee97d369100f86db3480ec84ff39aa0646a" exitCode=0 Nov 24 08:06:08 crc kubenswrapper[4799]: I1124 08:06:08.796819 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d14e-account-create-9lhsh" event={"ID":"508bd06c-03d1-44fe-9425-5a36f40a60b5","Type":"ContainerDied","Data":"cb9401eac7115b5bcb4c506c146c4ee97d369100f86db3480ec84ff39aa0646a"} Nov 24 08:06:08 crc kubenswrapper[4799]: I1124 08:06:08.796834 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d14e-account-create-9lhsh" event={"ID":"508bd06c-03d1-44fe-9425-5a36f40a60b5","Type":"ContainerStarted","Data":"f743fd2548ec3416cbcd22b97a82fc9aff3153e7f8d9ef2ca74db2b8d08307fe"} Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.206238 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dtcc5" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.217490 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d14e-account-create-9lhsh" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.395952 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/508bd06c-03d1-44fe-9425-5a36f40a60b5-operator-scripts\") pod \"508bd06c-03d1-44fe-9425-5a36f40a60b5\" (UID: \"508bd06c-03d1-44fe-9425-5a36f40a60b5\") " Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.396041 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzgrt\" (UniqueName: \"kubernetes.io/projected/508bd06c-03d1-44fe-9425-5a36f40a60b5-kube-api-access-hzgrt\") pod \"508bd06c-03d1-44fe-9425-5a36f40a60b5\" (UID: \"508bd06c-03d1-44fe-9425-5a36f40a60b5\") " Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.396189 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c5723e9-24ac-4cfa-b181-7030c18c2d27-operator-scripts\") pod \"4c5723e9-24ac-4cfa-b181-7030c18c2d27\" (UID: \"4c5723e9-24ac-4cfa-b181-7030c18c2d27\") " Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.396227 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pwns\" (UniqueName: \"kubernetes.io/projected/4c5723e9-24ac-4cfa-b181-7030c18c2d27-kube-api-access-4pwns\") pod \"4c5723e9-24ac-4cfa-b181-7030c18c2d27\" (UID: \"4c5723e9-24ac-4cfa-b181-7030c18c2d27\") " Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.396937 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c5723e9-24ac-4cfa-b181-7030c18c2d27-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4c5723e9-24ac-4cfa-b181-7030c18c2d27" (UID: "4c5723e9-24ac-4cfa-b181-7030c18c2d27"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.397135 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/508bd06c-03d1-44fe-9425-5a36f40a60b5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "508bd06c-03d1-44fe-9425-5a36f40a60b5" (UID: "508bd06c-03d1-44fe-9425-5a36f40a60b5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.402177 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c5723e9-24ac-4cfa-b181-7030c18c2d27-kube-api-access-4pwns" (OuterVolumeSpecName: "kube-api-access-4pwns") pod "4c5723e9-24ac-4cfa-b181-7030c18c2d27" (UID: "4c5723e9-24ac-4cfa-b181-7030c18c2d27"). InnerVolumeSpecName "kube-api-access-4pwns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.403069 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/508bd06c-03d1-44fe-9425-5a36f40a60b5-kube-api-access-hzgrt" (OuterVolumeSpecName: "kube-api-access-hzgrt") pod "508bd06c-03d1-44fe-9425-5a36f40a60b5" (UID: "508bd06c-03d1-44fe-9425-5a36f40a60b5"). InnerVolumeSpecName "kube-api-access-hzgrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.498572 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c5723e9-24ac-4cfa-b181-7030c18c2d27-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.498604 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pwns\" (UniqueName: \"kubernetes.io/projected/4c5723e9-24ac-4cfa-b181-7030c18c2d27-kube-api-access-4pwns\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.498615 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/508bd06c-03d1-44fe-9425-5a36f40a60b5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.498625 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzgrt\" (UniqueName: \"kubernetes.io/projected/508bd06c-03d1-44fe-9425-5a36f40a60b5-kube-api-access-hzgrt\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.817617 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dtcc5" event={"ID":"4c5723e9-24ac-4cfa-b181-7030c18c2d27","Type":"ContainerDied","Data":"eb62f4f4fe51de9e5aafe84de3985d3df4f757f36b31fbdaf15d5ce2ff10b413"} Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.817665 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dtcc5" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.817681 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb62f4f4fe51de9e5aafe84de3985d3df4f757f36b31fbdaf15d5ce2ff10b413" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.820205 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d14e-account-create-9lhsh" event={"ID":"508bd06c-03d1-44fe-9425-5a36f40a60b5","Type":"ContainerDied","Data":"f743fd2548ec3416cbcd22b97a82fc9aff3153e7f8d9ef2ca74db2b8d08307fe"} Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.820242 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f743fd2548ec3416cbcd22b97a82fc9aff3153e7f8d9ef2ca74db2b8d08307fe" Nov 24 08:06:10 crc kubenswrapper[4799]: I1124 08:06:10.820287 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d14e-account-create-9lhsh" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.556380 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-sw84r"] Nov 24 08:06:12 crc kubenswrapper[4799]: E1124 08:06:12.556728 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508bd06c-03d1-44fe-9425-5a36f40a60b5" containerName="mariadb-account-create" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.556740 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="508bd06c-03d1-44fe-9425-5a36f40a60b5" containerName="mariadb-account-create" Nov 24 08:06:12 crc kubenswrapper[4799]: E1124 08:06:12.556766 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5723e9-24ac-4cfa-b181-7030c18c2d27" containerName="mariadb-database-create" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.556772 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5723e9-24ac-4cfa-b181-7030c18c2d27" containerName="mariadb-database-create" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.556962 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="508bd06c-03d1-44fe-9425-5a36f40a60b5" containerName="mariadb-account-create" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.556976 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c5723e9-24ac-4cfa-b181-7030c18c2d27" containerName="mariadb-database-create" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.557494 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.563570 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.563575 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hlgpw" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.576696 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sw84r"] Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.634558 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-combined-ca-bundle\") pod \"barbican-db-sync-sw84r\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.634959 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdsfj\" (UniqueName: \"kubernetes.io/projected/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-kube-api-access-bdsfj\") pod \"barbican-db-sync-sw84r\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.634987 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-db-sync-config-data\") pod \"barbican-db-sync-sw84r\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.735896 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-combined-ca-bundle\") pod \"barbican-db-sync-sw84r\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.735956 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdsfj\" (UniqueName: \"kubernetes.io/projected/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-kube-api-access-bdsfj\") pod \"barbican-db-sync-sw84r\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.735976 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-db-sync-config-data\") pod \"barbican-db-sync-sw84r\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.740458 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-db-sync-config-data\") pod \"barbican-db-sync-sw84r\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.742764 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-combined-ca-bundle\") pod \"barbican-db-sync-sw84r\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.756772 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdsfj\" (UniqueName: \"kubernetes.io/projected/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-kube-api-access-bdsfj\") pod \"barbican-db-sync-sw84r\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:12 crc kubenswrapper[4799]: I1124 08:06:12.875691 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:13 crc kubenswrapper[4799]: I1124 08:06:13.310175 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sw84r"] Nov 24 08:06:13 crc kubenswrapper[4799]: I1124 08:06:13.847412 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sw84r" event={"ID":"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4","Type":"ContainerStarted","Data":"5eca0ca2d7420eca96af3fb0d6645c348eb9ce03aece37daea59106ec866d934"} Nov 24 08:06:17 crc kubenswrapper[4799]: I1124 08:06:17.891746 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sw84r" event={"ID":"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4","Type":"ContainerStarted","Data":"0c2ec8aeaf7aa6a29cd3ef59e6b1315cbc94672034fa02675d5e4275282a8156"} Nov 24 08:06:17 crc kubenswrapper[4799]: I1124 08:06:17.921046 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-sw84r" podStartSLOduration=1.8690825850000001 podStartE2EDuration="5.921022623s" podCreationTimestamp="2025-11-24 08:06:12 +0000 UTC" firstStartedPulling="2025-11-24 08:06:13.33008798 +0000 UTC m=+4718.986070454" lastFinishedPulling="2025-11-24 08:06:17.382027988 +0000 UTC m=+4723.038010492" observedRunningTime="2025-11-24 08:06:17.905232094 +0000 UTC m=+4723.561214578" watchObservedRunningTime="2025-11-24 08:06:17.921022623 +0000 UTC m=+4723.577005117" Nov 24 08:06:19 crc kubenswrapper[4799]: I1124 08:06:19.912165 4799 generic.go:334] "Generic (PLEG): container finished" podID="d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4" containerID="0c2ec8aeaf7aa6a29cd3ef59e6b1315cbc94672034fa02675d5e4275282a8156" exitCode=0 Nov 24 08:06:19 crc kubenswrapper[4799]: I1124 08:06:19.912314 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sw84r" event={"ID":"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4","Type":"ContainerDied","Data":"0c2ec8aeaf7aa6a29cd3ef59e6b1315cbc94672034fa02675d5e4275282a8156"} Nov 24 08:06:20 crc kubenswrapper[4799]: I1124 08:06:20.401056 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:06:20 crc kubenswrapper[4799]: I1124 08:06:20.401147 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.244960 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.389677 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-db-sync-config-data\") pod \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.389767 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-combined-ca-bundle\") pod \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.389809 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdsfj\" (UniqueName: \"kubernetes.io/projected/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-kube-api-access-bdsfj\") pod \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\" (UID: \"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4\") " Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.396603 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-kube-api-access-bdsfj" (OuterVolumeSpecName: "kube-api-access-bdsfj") pod "d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4" (UID: "d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4"). InnerVolumeSpecName "kube-api-access-bdsfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.396715 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4" (UID: "d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.428151 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4" (UID: "d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.491279 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.491313 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdsfj\" (UniqueName: \"kubernetes.io/projected/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-kube-api-access-bdsfj\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.491326 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.933382 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sw84r" event={"ID":"d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4","Type":"ContainerDied","Data":"5eca0ca2d7420eca96af3fb0d6645c348eb9ce03aece37daea59106ec866d934"} Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.933734 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eca0ca2d7420eca96af3fb0d6645c348eb9ce03aece37daea59106ec866d934" Nov 24 08:06:21 crc kubenswrapper[4799]: I1124 08:06:21.933435 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sw84r" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.180880 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-58fd65459-m8dgs"] Nov 24 08:06:22 crc kubenswrapper[4799]: E1124 08:06:22.181223 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4" containerName="barbican-db-sync" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.181238 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4" containerName="barbican-db-sync" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.181405 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4" containerName="barbican-db-sync" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.184874 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.189823 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.189994 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hlgpw" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.191381 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.214177 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-58fd65459-m8dgs"] Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.227980 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6fb47964bb-kqvtg"] Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.229351 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.232905 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.257629 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6fb47964bb-kqvtg"] Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.271920 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68d59b85bf-kfkc9"] Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.273342 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.280705 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d59b85bf-kfkc9"] Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.303518 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e93b600f-f9b4-4bda-88a0-86713f38ef9e-config-data\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.303574 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8cw7\" (UniqueName: \"kubernetes.io/projected/e93b600f-f9b4-4bda-88a0-86713f38ef9e-kube-api-access-k8cw7\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.303604 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e93b600f-f9b4-4bda-88a0-86713f38ef9e-logs\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.303644 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e93b600f-f9b4-4bda-88a0-86713f38ef9e-combined-ca-bundle\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.303681 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e93b600f-f9b4-4bda-88a0-86713f38ef9e-config-data-custom\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.341951 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-85f6f78554-znl22"] Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.343776 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.357484 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.364497 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85f6f78554-znl22"] Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406151 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54a286e6-dbd2-4423-b15b-6317aec5119c-config-data-custom\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406196 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx5tj\" (UniqueName: \"kubernetes.io/projected/54a286e6-dbd2-4423-b15b-6317aec5119c-kube-api-access-rx5tj\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406220 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-dns-svc\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406242 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a286e6-dbd2-4423-b15b-6317aec5119c-config-data\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406279 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-config\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406303 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e93b600f-f9b4-4bda-88a0-86713f38ef9e-config-data\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406333 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8cw7\" (UniqueName: \"kubernetes.io/projected/e93b600f-f9b4-4bda-88a0-86713f38ef9e-kube-api-access-k8cw7\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406349 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blx8j\" (UniqueName: \"kubernetes.io/projected/375fe805-7346-48ac-a330-f28ff5da9303-kube-api-access-blx8j\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406368 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a286e6-dbd2-4423-b15b-6317aec5119c-combined-ca-bundle\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406388 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e93b600f-f9b4-4bda-88a0-86713f38ef9e-logs\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406408 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-sb\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406435 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54a286e6-dbd2-4423-b15b-6317aec5119c-logs\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406458 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e93b600f-f9b4-4bda-88a0-86713f38ef9e-combined-ca-bundle\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406495 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e93b600f-f9b4-4bda-88a0-86713f38ef9e-config-data-custom\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.406518 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-nb\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.407171 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e93b600f-f9b4-4bda-88a0-86713f38ef9e-logs\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508098 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-sb\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508180 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54a286e6-dbd2-4423-b15b-6317aec5119c-logs\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508216 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f2d5e05-6093-4786-8940-c5039d601979-config-data\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508303 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-nb\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508330 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54a286e6-dbd2-4423-b15b-6317aec5119c-config-data-custom\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508369 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx5tj\" (UniqueName: \"kubernetes.io/projected/54a286e6-dbd2-4423-b15b-6317aec5119c-kube-api-access-rx5tj\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508387 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f2d5e05-6093-4786-8940-c5039d601979-logs\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508411 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-dns-svc\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508435 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a286e6-dbd2-4423-b15b-6317aec5119c-config-data\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508458 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f2d5e05-6093-4786-8940-c5039d601979-combined-ca-bundle\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508488 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z6dq\" (UniqueName: \"kubernetes.io/projected/4f2d5e05-6093-4786-8940-c5039d601979-kube-api-access-2z6dq\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508508 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-config\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508548 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f2d5e05-6093-4786-8940-c5039d601979-config-data-custom\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508576 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blx8j\" (UniqueName: \"kubernetes.io/projected/375fe805-7346-48ac-a330-f28ff5da9303-kube-api-access-blx8j\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.508596 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a286e6-dbd2-4423-b15b-6317aec5119c-combined-ca-bundle\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.509407 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-sb\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.509469 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-nb\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.509773 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-config\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.509801 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54a286e6-dbd2-4423-b15b-6317aec5119c-logs\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.510123 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-dns-svc\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.610121 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f2d5e05-6093-4786-8940-c5039d601979-logs\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.610201 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f2d5e05-6093-4786-8940-c5039d601979-combined-ca-bundle\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.610240 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z6dq\" (UniqueName: \"kubernetes.io/projected/4f2d5e05-6093-4786-8940-c5039d601979-kube-api-access-2z6dq\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.610286 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f2d5e05-6093-4786-8940-c5039d601979-config-data-custom\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.610368 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f2d5e05-6093-4786-8940-c5039d601979-config-data\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.610692 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f2d5e05-6093-4786-8940-c5039d601979-logs\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.675165 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e93b600f-f9b4-4bda-88a0-86713f38ef9e-config-data-custom\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.675489 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e93b600f-f9b4-4bda-88a0-86713f38ef9e-config-data\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.675652 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e93b600f-f9b4-4bda-88a0-86713f38ef9e-combined-ca-bundle\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.675762 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a286e6-dbd2-4423-b15b-6317aec5119c-combined-ca-bundle\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.675971 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8cw7\" (UniqueName: \"kubernetes.io/projected/e93b600f-f9b4-4bda-88a0-86713f38ef9e-kube-api-access-k8cw7\") pod \"barbican-worker-58fd65459-m8dgs\" (UID: \"e93b600f-f9b4-4bda-88a0-86713f38ef9e\") " pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.677545 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54a286e6-dbd2-4423-b15b-6317aec5119c-config-data-custom\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.677755 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx5tj\" (UniqueName: \"kubernetes.io/projected/54a286e6-dbd2-4423-b15b-6317aec5119c-kube-api-access-rx5tj\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.678169 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blx8j\" (UniqueName: \"kubernetes.io/projected/375fe805-7346-48ac-a330-f28ff5da9303-kube-api-access-blx8j\") pod \"dnsmasq-dns-68d59b85bf-kfkc9\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.678329 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a286e6-dbd2-4423-b15b-6317aec5119c-config-data\") pod \"barbican-keystone-listener-6fb47964bb-kqvtg\" (UID: \"54a286e6-dbd2-4423-b15b-6317aec5119c\") " pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.679063 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f2d5e05-6093-4786-8940-c5039d601979-config-data-custom\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.679550 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f2d5e05-6093-4786-8940-c5039d601979-config-data\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.679551 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f2d5e05-6093-4786-8940-c5039d601979-combined-ca-bundle\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.679989 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z6dq\" (UniqueName: \"kubernetes.io/projected/4f2d5e05-6093-4786-8940-c5039d601979-kube-api-access-2z6dq\") pod \"barbican-api-85f6f78554-znl22\" (UID: \"4f2d5e05-6093-4786-8940-c5039d601979\") " pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.828064 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-58fd65459-m8dgs" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.867402 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.892021 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:22 crc kubenswrapper[4799]: I1124 08:06:22.971992 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:23 crc kubenswrapper[4799]: I1124 08:06:23.346039 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-58fd65459-m8dgs"] Nov 24 08:06:23 crc kubenswrapper[4799]: I1124 08:06:23.377020 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6fb47964bb-kqvtg"] Nov 24 08:06:23 crc kubenswrapper[4799]: I1124 08:06:23.450142 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d59b85bf-kfkc9"] Nov 24 08:06:23 crc kubenswrapper[4799]: W1124 08:06:23.457648 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod375fe805_7346_48ac_a330_f28ff5da9303.slice/crio-b82fdc447f782bd137e330ec34d98b9209aaf5a6f36cd6372346829468f407d9 WatchSource:0}: Error finding container b82fdc447f782bd137e330ec34d98b9209aaf5a6f36cd6372346829468f407d9: Status 404 returned error can't find the container with id b82fdc447f782bd137e330ec34d98b9209aaf5a6f36cd6372346829468f407d9 Nov 24 08:06:23 crc kubenswrapper[4799]: I1124 08:06:23.569107 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85f6f78554-znl22"] Nov 24 08:06:23 crc kubenswrapper[4799]: W1124 08:06:23.571459 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f2d5e05_6093_4786_8940_c5039d601979.slice/crio-205f02c741cc82eeec5053aa4bf5cdd97a08456a4c0fd8e6880f2b21bf3064e1 WatchSource:0}: Error finding container 205f02c741cc82eeec5053aa4bf5cdd97a08456a4c0fd8e6880f2b21bf3064e1: Status 404 returned error can't find the container with id 205f02c741cc82eeec5053aa4bf5cdd97a08456a4c0fd8e6880f2b21bf3064e1 Nov 24 08:06:23 crc kubenswrapper[4799]: I1124 08:06:23.966505 4799 generic.go:334] "Generic (PLEG): container finished" podID="375fe805-7346-48ac-a330-f28ff5da9303" containerID="0c92f88b54076937c229edca66bebf876ca670b468c18ca083e68a776ef22cfc" exitCode=0 Nov 24 08:06:23 crc kubenswrapper[4799]: I1124 08:06:23.966703 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" event={"ID":"375fe805-7346-48ac-a330-f28ff5da9303","Type":"ContainerDied","Data":"0c92f88b54076937c229edca66bebf876ca670b468c18ca083e68a776ef22cfc"} Nov 24 08:06:23 crc kubenswrapper[4799]: I1124 08:06:23.968266 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" event={"ID":"375fe805-7346-48ac-a330-f28ff5da9303","Type":"ContainerStarted","Data":"b82fdc447f782bd137e330ec34d98b9209aaf5a6f36cd6372346829468f407d9"} Nov 24 08:06:23 crc kubenswrapper[4799]: I1124 08:06:23.971741 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-58fd65459-m8dgs" event={"ID":"e93b600f-f9b4-4bda-88a0-86713f38ef9e","Type":"ContainerStarted","Data":"cc63fdfed70cbd875522c71d9fc62784134df27c31eb48f80e2a1b5d6a605f5b"} Nov 24 08:06:23 crc kubenswrapper[4799]: I1124 08:06:23.974299 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" event={"ID":"54a286e6-dbd2-4423-b15b-6317aec5119c","Type":"ContainerStarted","Data":"61e7c4dc9a71d6a6cd02398655a21ce3a91eb4df5fedb4bb878bb32921d3d47a"} Nov 24 08:06:23 crc kubenswrapper[4799]: I1124 08:06:23.976156 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85f6f78554-znl22" event={"ID":"4f2d5e05-6093-4786-8940-c5039d601979","Type":"ContainerStarted","Data":"878c070eb689371bb9c6605041968c9a1d30493f6a30ba1c5362baa622080a8f"} Nov 24 08:06:23 crc kubenswrapper[4799]: I1124 08:06:23.976193 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85f6f78554-znl22" event={"ID":"4f2d5e05-6093-4786-8940-c5039d601979","Type":"ContainerStarted","Data":"205f02c741cc82eeec5053aa4bf5cdd97a08456a4c0fd8e6880f2b21bf3064e1"} Nov 24 08:06:24 crc kubenswrapper[4799]: I1124 08:06:24.990296 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" event={"ID":"375fe805-7346-48ac-a330-f28ff5da9303","Type":"ContainerStarted","Data":"065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4"} Nov 24 08:06:24 crc kubenswrapper[4799]: I1124 08:06:24.990740 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:24 crc kubenswrapper[4799]: I1124 08:06:24.995141 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85f6f78554-znl22" event={"ID":"4f2d5e05-6093-4786-8940-c5039d601979","Type":"ContainerStarted","Data":"9e48c44039c7485c16397e8250dc7d97a073b71e5b413761407e5702c3131435"} Nov 24 08:06:24 crc kubenswrapper[4799]: I1124 08:06:24.995242 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:24 crc kubenswrapper[4799]: I1124 08:06:24.995289 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:25 crc kubenswrapper[4799]: I1124 08:06:25.014013 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" podStartSLOduration=3.013980217 podStartE2EDuration="3.013980217s" podCreationTimestamp="2025-11-24 08:06:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:06:25.006792977 +0000 UTC m=+4730.662775441" watchObservedRunningTime="2025-11-24 08:06:25.013980217 +0000 UTC m=+4730.669962701" Nov 24 08:06:25 crc kubenswrapper[4799]: I1124 08:06:25.041528 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-85f6f78554-znl22" podStartSLOduration=3.041289685 podStartE2EDuration="3.041289685s" podCreationTimestamp="2025-11-24 08:06:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:06:25.034385004 +0000 UTC m=+4730.690367498" watchObservedRunningTime="2025-11-24 08:06:25.041289685 +0000 UTC m=+4730.697272199" Nov 24 08:06:26 crc kubenswrapper[4799]: I1124 08:06:26.008930 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" event={"ID":"54a286e6-dbd2-4423-b15b-6317aec5119c","Type":"ContainerStarted","Data":"3f5d793c5fd73c8e7381d43b36c5ab1b705b021ad8688266a429e0bc947c8348"} Nov 24 08:06:26 crc kubenswrapper[4799]: I1124 08:06:26.009283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" event={"ID":"54a286e6-dbd2-4423-b15b-6317aec5119c","Type":"ContainerStarted","Data":"3e19ce3cdd837a22825aa94fdcc6b9ce53ed390e53bfa68a73107e4493217b4e"} Nov 24 08:06:26 crc kubenswrapper[4799]: I1124 08:06:26.013678 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-58fd65459-m8dgs" event={"ID":"e93b600f-f9b4-4bda-88a0-86713f38ef9e","Type":"ContainerStarted","Data":"4d20cc7bb1e2594f7b8c18cc02b73e9f88e289eaee916c79b4c3adae403edea5"} Nov 24 08:06:26 crc kubenswrapper[4799]: I1124 08:06:26.013819 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-58fd65459-m8dgs" event={"ID":"e93b600f-f9b4-4bda-88a0-86713f38ef9e","Type":"ContainerStarted","Data":"bca360800975d58f653e3fde8d4e63d20907b30228719edc7d79515293000600"} Nov 24 08:06:26 crc kubenswrapper[4799]: I1124 08:06:26.039347 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6fb47964bb-kqvtg" podStartSLOduration=2.3044910339999998 podStartE2EDuration="4.039322275s" podCreationTimestamp="2025-11-24 08:06:22 +0000 UTC" firstStartedPulling="2025-11-24 08:06:23.377964141 +0000 UTC m=+4729.033946615" lastFinishedPulling="2025-11-24 08:06:25.112795372 +0000 UTC m=+4730.768777856" observedRunningTime="2025-11-24 08:06:26.029164573 +0000 UTC m=+4731.685147087" watchObservedRunningTime="2025-11-24 08:06:26.039322275 +0000 UTC m=+4731.695304759" Nov 24 08:06:26 crc kubenswrapper[4799]: I1124 08:06:26.060158 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-58fd65459-m8dgs" podStartSLOduration=2.31513709 podStartE2EDuration="4.060127683s" podCreationTimestamp="2025-11-24 08:06:22 +0000 UTC" firstStartedPulling="2025-11-24 08:06:23.371148912 +0000 UTC m=+4729.027131386" lastFinishedPulling="2025-11-24 08:06:25.116139505 +0000 UTC m=+4730.772121979" observedRunningTime="2025-11-24 08:06:26.057935653 +0000 UTC m=+4731.713918167" watchObservedRunningTime="2025-11-24 08:06:26.060127683 +0000 UTC m=+4731.716110187" Nov 24 08:06:32 crc kubenswrapper[4799]: I1124 08:06:32.895066 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:06:32 crc kubenswrapper[4799]: I1124 08:06:32.976103 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f4455545-ch4nv"] Nov 24 08:06:32 crc kubenswrapper[4799]: I1124 08:06:32.976398 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" podUID="482dfc9d-977f-47a3-85b0-4722749aa333" containerName="dnsmasq-dns" containerID="cri-o://fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7" gracePeriod=10 Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.676923 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.826818 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-config\") pod \"482dfc9d-977f-47a3-85b0-4722749aa333\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.827153 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-dns-svc\") pod \"482dfc9d-977f-47a3-85b0-4722749aa333\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.827218 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-sb\") pod \"482dfc9d-977f-47a3-85b0-4722749aa333\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.827331 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnrrp\" (UniqueName: \"kubernetes.io/projected/482dfc9d-977f-47a3-85b0-4722749aa333-kube-api-access-wnrrp\") pod \"482dfc9d-977f-47a3-85b0-4722749aa333\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.827417 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-nb\") pod \"482dfc9d-977f-47a3-85b0-4722749aa333\" (UID: \"482dfc9d-977f-47a3-85b0-4722749aa333\") " Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.848599 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/482dfc9d-977f-47a3-85b0-4722749aa333-kube-api-access-wnrrp" (OuterVolumeSpecName: "kube-api-access-wnrrp") pod "482dfc9d-977f-47a3-85b0-4722749aa333" (UID: "482dfc9d-977f-47a3-85b0-4722749aa333"). InnerVolumeSpecName "kube-api-access-wnrrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.871509 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "482dfc9d-977f-47a3-85b0-4722749aa333" (UID: "482dfc9d-977f-47a3-85b0-4722749aa333"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.885909 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-config" (OuterVolumeSpecName: "config") pod "482dfc9d-977f-47a3-85b0-4722749aa333" (UID: "482dfc9d-977f-47a3-85b0-4722749aa333"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.888737 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "482dfc9d-977f-47a3-85b0-4722749aa333" (UID: "482dfc9d-977f-47a3-85b0-4722749aa333"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.894306 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "482dfc9d-977f-47a3-85b0-4722749aa333" (UID: "482dfc9d-977f-47a3-85b0-4722749aa333"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.929576 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.929612 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnrrp\" (UniqueName: \"kubernetes.io/projected/482dfc9d-977f-47a3-85b0-4722749aa333-kube-api-access-wnrrp\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.929623 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.929633 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:33 crc kubenswrapper[4799]: I1124 08:06:33.929642 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/482dfc9d-977f-47a3-85b0-4722749aa333-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.081010 4799 generic.go:334] "Generic (PLEG): container finished" podID="482dfc9d-977f-47a3-85b0-4722749aa333" containerID="fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7" exitCode=0 Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.081080 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" event={"ID":"482dfc9d-977f-47a3-85b0-4722749aa333","Type":"ContainerDied","Data":"fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7"} Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.081129 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" event={"ID":"482dfc9d-977f-47a3-85b0-4722749aa333","Type":"ContainerDied","Data":"ae44798e80098b2260de93536a1b2d842bc37a51e2c0e1e445a9064dd8c2bf99"} Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.081148 4799 scope.go:117] "RemoveContainer" containerID="fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7" Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.081320 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f4455545-ch4nv" Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.151478 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f4455545-ch4nv"] Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.160350 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f4455545-ch4nv"] Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.163912 4799 scope.go:117] "RemoveContainer" containerID="098fdf7b84a004818b6be1a96bf12d99ba29aa87f9a5ed15d708ce9751730c69" Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.196195 4799 scope.go:117] "RemoveContainer" containerID="fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7" Nov 24 08:06:34 crc kubenswrapper[4799]: E1124 08:06:34.197648 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7\": container with ID starting with fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7 not found: ID does not exist" containerID="fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7" Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.197689 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7"} err="failed to get container status \"fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7\": rpc error: code = NotFound desc = could not find container \"fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7\": container with ID starting with fcc77e41638542945c7a1172961648e856a39387d5a4c623306174b5589faed7 not found: ID does not exist" Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.197720 4799 scope.go:117] "RemoveContainer" containerID="098fdf7b84a004818b6be1a96bf12d99ba29aa87f9a5ed15d708ce9751730c69" Nov 24 08:06:34 crc kubenswrapper[4799]: E1124 08:06:34.198035 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"098fdf7b84a004818b6be1a96bf12d99ba29aa87f9a5ed15d708ce9751730c69\": container with ID starting with 098fdf7b84a004818b6be1a96bf12d99ba29aa87f9a5ed15d708ce9751730c69 not found: ID does not exist" containerID="098fdf7b84a004818b6be1a96bf12d99ba29aa87f9a5ed15d708ce9751730c69" Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.198061 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"098fdf7b84a004818b6be1a96bf12d99ba29aa87f9a5ed15d708ce9751730c69"} err="failed to get container status \"098fdf7b84a004818b6be1a96bf12d99ba29aa87f9a5ed15d708ce9751730c69\": rpc error: code = NotFound desc = could not find container \"098fdf7b84a004818b6be1a96bf12d99ba29aa87f9a5ed15d708ce9751730c69\": container with ID starting with 098fdf7b84a004818b6be1a96bf12d99ba29aa87f9a5ed15d708ce9751730c69 not found: ID does not exist" Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.572807 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:34 crc kubenswrapper[4799]: I1124 08:06:34.618167 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85f6f78554-znl22" Nov 24 08:06:35 crc kubenswrapper[4799]: I1124 08:06:35.639232 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="482dfc9d-977f-47a3-85b0-4722749aa333" path="/var/lib/kubelet/pods/482dfc9d-977f-47a3-85b0-4722749aa333/volumes" Nov 24 08:06:40 crc kubenswrapper[4799]: E1124 08:06:40.181561 4799 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:59656->38.102.83.50:45263: write tcp 38.102.83.50:59656->38.102.83.50:45263: write: broken pipe Nov 24 08:06:43 crc kubenswrapper[4799]: E1124 08:06:43.200704 4799 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:49244->38.102.83.50:45263: write tcp 38.102.83.50:49244->38.102.83.50:45263: write: broken pipe Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.203491 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-lpbqc"] Nov 24 08:06:46 crc kubenswrapper[4799]: E1124 08:06:46.204281 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482dfc9d-977f-47a3-85b0-4722749aa333" containerName="dnsmasq-dns" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.204294 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="482dfc9d-977f-47a3-85b0-4722749aa333" containerName="dnsmasq-dns" Nov 24 08:06:46 crc kubenswrapper[4799]: E1124 08:06:46.204320 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482dfc9d-977f-47a3-85b0-4722749aa333" containerName="init" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.204326 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="482dfc9d-977f-47a3-85b0-4722749aa333" containerName="init" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.204489 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="482dfc9d-977f-47a3-85b0-4722749aa333" containerName="dnsmasq-dns" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.205010 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lpbqc" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.215770 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lpbqc"] Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.305991 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-3e6b-account-create-48czt"] Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.307094 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e6b-account-create-48czt" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.309471 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.321329 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3e6b-account-create-48czt"] Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.371980 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr6tz\" (UniqueName: \"kubernetes.io/projected/96d54256-8e3b-4ccb-96b5-39bad6ebf371-kube-api-access-wr6tz\") pod \"neutron-db-create-lpbqc\" (UID: \"96d54256-8e3b-4ccb-96b5-39bad6ebf371\") " pod="openstack/neutron-db-create-lpbqc" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.372049 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96d54256-8e3b-4ccb-96b5-39bad6ebf371-operator-scripts\") pod \"neutron-db-create-lpbqc\" (UID: \"96d54256-8e3b-4ccb-96b5-39bad6ebf371\") " pod="openstack/neutron-db-create-lpbqc" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.475831 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr6tz\" (UniqueName: \"kubernetes.io/projected/96d54256-8e3b-4ccb-96b5-39bad6ebf371-kube-api-access-wr6tz\") pod \"neutron-db-create-lpbqc\" (UID: \"96d54256-8e3b-4ccb-96b5-39bad6ebf371\") " pod="openstack/neutron-db-create-lpbqc" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.475910 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96d54256-8e3b-4ccb-96b5-39bad6ebf371-operator-scripts\") pod \"neutron-db-create-lpbqc\" (UID: \"96d54256-8e3b-4ccb-96b5-39bad6ebf371\") " pod="openstack/neutron-db-create-lpbqc" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.475971 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-operator-scripts\") pod \"neutron-3e6b-account-create-48czt\" (UID: \"3de7dad8-7eca-4d02-ac2f-8d07a3608d75\") " pod="openstack/neutron-3e6b-account-create-48czt" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.476038 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzs6l\" (UniqueName: \"kubernetes.io/projected/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-kube-api-access-tzs6l\") pod \"neutron-3e6b-account-create-48czt\" (UID: \"3de7dad8-7eca-4d02-ac2f-8d07a3608d75\") " pod="openstack/neutron-3e6b-account-create-48czt" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.476814 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96d54256-8e3b-4ccb-96b5-39bad6ebf371-operator-scripts\") pod \"neutron-db-create-lpbqc\" (UID: \"96d54256-8e3b-4ccb-96b5-39bad6ebf371\") " pod="openstack/neutron-db-create-lpbqc" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.495898 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr6tz\" (UniqueName: \"kubernetes.io/projected/96d54256-8e3b-4ccb-96b5-39bad6ebf371-kube-api-access-wr6tz\") pod \"neutron-db-create-lpbqc\" (UID: \"96d54256-8e3b-4ccb-96b5-39bad6ebf371\") " pod="openstack/neutron-db-create-lpbqc" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.521727 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lpbqc" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.577075 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-operator-scripts\") pod \"neutron-3e6b-account-create-48czt\" (UID: \"3de7dad8-7eca-4d02-ac2f-8d07a3608d75\") " pod="openstack/neutron-3e6b-account-create-48czt" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.577145 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzs6l\" (UniqueName: \"kubernetes.io/projected/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-kube-api-access-tzs6l\") pod \"neutron-3e6b-account-create-48czt\" (UID: \"3de7dad8-7eca-4d02-ac2f-8d07a3608d75\") " pod="openstack/neutron-3e6b-account-create-48czt" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.578060 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-operator-scripts\") pod \"neutron-3e6b-account-create-48czt\" (UID: \"3de7dad8-7eca-4d02-ac2f-8d07a3608d75\") " pod="openstack/neutron-3e6b-account-create-48czt" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.598426 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzs6l\" (UniqueName: \"kubernetes.io/projected/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-kube-api-access-tzs6l\") pod \"neutron-3e6b-account-create-48czt\" (UID: \"3de7dad8-7eca-4d02-ac2f-8d07a3608d75\") " pod="openstack/neutron-3e6b-account-create-48czt" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.625247 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e6b-account-create-48czt" Nov 24 08:06:46 crc kubenswrapper[4799]: I1124 08:06:46.992429 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lpbqc"] Nov 24 08:06:47 crc kubenswrapper[4799]: I1124 08:06:47.128139 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3e6b-account-create-48czt"] Nov 24 08:06:47 crc kubenswrapper[4799]: W1124 08:06:47.133190 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3de7dad8_7eca_4d02_ac2f_8d07a3608d75.slice/crio-c9bc4b4acb16a6fe1c011584f3a3862d8ab33f279647d4874f50e6990eaf3eca WatchSource:0}: Error finding container c9bc4b4acb16a6fe1c011584f3a3862d8ab33f279647d4874f50e6990eaf3eca: Status 404 returned error can't find the container with id c9bc4b4acb16a6fe1c011584f3a3862d8ab33f279647d4874f50e6990eaf3eca Nov 24 08:06:47 crc kubenswrapper[4799]: I1124 08:06:47.212551 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3e6b-account-create-48czt" event={"ID":"3de7dad8-7eca-4d02-ac2f-8d07a3608d75","Type":"ContainerStarted","Data":"c9bc4b4acb16a6fe1c011584f3a3862d8ab33f279647d4874f50e6990eaf3eca"} Nov 24 08:06:47 crc kubenswrapper[4799]: I1124 08:06:47.214976 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lpbqc" event={"ID":"96d54256-8e3b-4ccb-96b5-39bad6ebf371","Type":"ContainerStarted","Data":"62da8edc3ccefd69dbdf75836c637b51504f6e77efd89d463a51555183fdb306"} Nov 24 08:06:48 crc kubenswrapper[4799]: I1124 08:06:48.227107 4799 generic.go:334] "Generic (PLEG): container finished" podID="3de7dad8-7eca-4d02-ac2f-8d07a3608d75" containerID="01ea2798e6be3b95b69f465afe110ac272165a163aeeed1ee26cfa6510047bd6" exitCode=0 Nov 24 08:06:48 crc kubenswrapper[4799]: I1124 08:06:48.227197 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3e6b-account-create-48czt" event={"ID":"3de7dad8-7eca-4d02-ac2f-8d07a3608d75","Type":"ContainerDied","Data":"01ea2798e6be3b95b69f465afe110ac272165a163aeeed1ee26cfa6510047bd6"} Nov 24 08:06:48 crc kubenswrapper[4799]: I1124 08:06:48.230015 4799 generic.go:334] "Generic (PLEG): container finished" podID="96d54256-8e3b-4ccb-96b5-39bad6ebf371" containerID="38a0f8c178a84731da58580d7dfa58bef30c8740c5c85746ae760163e298b8d4" exitCode=0 Nov 24 08:06:48 crc kubenswrapper[4799]: I1124 08:06:48.230075 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lpbqc" event={"ID":"96d54256-8e3b-4ccb-96b5-39bad6ebf371","Type":"ContainerDied","Data":"38a0f8c178a84731da58580d7dfa58bef30c8740c5c85746ae760163e298b8d4"} Nov 24 08:06:49 crc kubenswrapper[4799]: I1124 08:06:49.744980 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lpbqc" Nov 24 08:06:49 crc kubenswrapper[4799]: I1124 08:06:49.750196 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96d54256-8e3b-4ccb-96b5-39bad6ebf371-operator-scripts\") pod \"96d54256-8e3b-4ccb-96b5-39bad6ebf371\" (UID: \"96d54256-8e3b-4ccb-96b5-39bad6ebf371\") " Nov 24 08:06:49 crc kubenswrapper[4799]: I1124 08:06:49.750258 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr6tz\" (UniqueName: \"kubernetes.io/projected/96d54256-8e3b-4ccb-96b5-39bad6ebf371-kube-api-access-wr6tz\") pod \"96d54256-8e3b-4ccb-96b5-39bad6ebf371\" (UID: \"96d54256-8e3b-4ccb-96b5-39bad6ebf371\") " Nov 24 08:06:49 crc kubenswrapper[4799]: I1124 08:06:49.751662 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96d54256-8e3b-4ccb-96b5-39bad6ebf371-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "96d54256-8e3b-4ccb-96b5-39bad6ebf371" (UID: "96d54256-8e3b-4ccb-96b5-39bad6ebf371"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:06:49 crc kubenswrapper[4799]: I1124 08:06:49.768647 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96d54256-8e3b-4ccb-96b5-39bad6ebf371-kube-api-access-wr6tz" (OuterVolumeSpecName: "kube-api-access-wr6tz") pod "96d54256-8e3b-4ccb-96b5-39bad6ebf371" (UID: "96d54256-8e3b-4ccb-96b5-39bad6ebf371"). InnerVolumeSpecName "kube-api-access-wr6tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:06:49 crc kubenswrapper[4799]: I1124 08:06:49.851937 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96d54256-8e3b-4ccb-96b5-39bad6ebf371-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:49 crc kubenswrapper[4799]: I1124 08:06:49.851974 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr6tz\" (UniqueName: \"kubernetes.io/projected/96d54256-8e3b-4ccb-96b5-39bad6ebf371-kube-api-access-wr6tz\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:49 crc kubenswrapper[4799]: I1124 08:06:49.863562 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e6b-account-create-48czt" Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.055492 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-operator-scripts\") pod \"3de7dad8-7eca-4d02-ac2f-8d07a3608d75\" (UID: \"3de7dad8-7eca-4d02-ac2f-8d07a3608d75\") " Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.055782 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzs6l\" (UniqueName: \"kubernetes.io/projected/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-kube-api-access-tzs6l\") pod \"3de7dad8-7eca-4d02-ac2f-8d07a3608d75\" (UID: \"3de7dad8-7eca-4d02-ac2f-8d07a3608d75\") " Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.056259 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3de7dad8-7eca-4d02-ac2f-8d07a3608d75" (UID: "3de7dad8-7eca-4d02-ac2f-8d07a3608d75"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.056669 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.059187 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-kube-api-access-tzs6l" (OuterVolumeSpecName: "kube-api-access-tzs6l") pod "3de7dad8-7eca-4d02-ac2f-8d07a3608d75" (UID: "3de7dad8-7eca-4d02-ac2f-8d07a3608d75"). InnerVolumeSpecName "kube-api-access-tzs6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.158414 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzs6l\" (UniqueName: \"kubernetes.io/projected/3de7dad8-7eca-4d02-ac2f-8d07a3608d75-kube-api-access-tzs6l\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.248697 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3e6b-account-create-48czt" event={"ID":"3de7dad8-7eca-4d02-ac2f-8d07a3608d75","Type":"ContainerDied","Data":"c9bc4b4acb16a6fe1c011584f3a3862d8ab33f279647d4874f50e6990eaf3eca"} Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.248763 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9bc4b4acb16a6fe1c011584f3a3862d8ab33f279647d4874f50e6990eaf3eca" Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.248866 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e6b-account-create-48czt" Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.251493 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lpbqc" event={"ID":"96d54256-8e3b-4ccb-96b5-39bad6ebf371","Type":"ContainerDied","Data":"62da8edc3ccefd69dbdf75836c637b51504f6e77efd89d463a51555183fdb306"} Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.251617 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62da8edc3ccefd69dbdf75836c637b51504f6e77efd89d463a51555183fdb306" Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.251536 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lpbqc" Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.400323 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:06:50 crc kubenswrapper[4799]: I1124 08:06:50.400405 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.507048 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-kcww7"] Nov 24 08:06:51 crc kubenswrapper[4799]: E1124 08:06:51.507788 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3de7dad8-7eca-4d02-ac2f-8d07a3608d75" containerName="mariadb-account-create" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.507806 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3de7dad8-7eca-4d02-ac2f-8d07a3608d75" containerName="mariadb-account-create" Nov 24 08:06:51 crc kubenswrapper[4799]: E1124 08:06:51.507825 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96d54256-8e3b-4ccb-96b5-39bad6ebf371" containerName="mariadb-database-create" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.507834 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="96d54256-8e3b-4ccb-96b5-39bad6ebf371" containerName="mariadb-database-create" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.508100 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="96d54256-8e3b-4ccb-96b5-39bad6ebf371" containerName="mariadb-database-create" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.508143 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3de7dad8-7eca-4d02-ac2f-8d07a3608d75" containerName="mariadb-account-create" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.508816 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.512311 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.512429 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dxmcv" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.512928 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.517784 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kcww7"] Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.685578 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-combined-ca-bundle\") pod \"neutron-db-sync-kcww7\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.686017 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-config\") pod \"neutron-db-sync-kcww7\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.686215 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbztd\" (UniqueName: \"kubernetes.io/projected/8969b312-9957-41a5-a640-635f79eb16e4-kube-api-access-gbztd\") pod \"neutron-db-sync-kcww7\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.787375 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-config\") pod \"neutron-db-sync-kcww7\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.787687 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbztd\" (UniqueName: \"kubernetes.io/projected/8969b312-9957-41a5-a640-635f79eb16e4-kube-api-access-gbztd\") pod \"neutron-db-sync-kcww7\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.787810 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-combined-ca-bundle\") pod \"neutron-db-sync-kcww7\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.792185 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-config\") pod \"neutron-db-sync-kcww7\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.792730 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-combined-ca-bundle\") pod \"neutron-db-sync-kcww7\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.806691 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbztd\" (UniqueName: \"kubernetes.io/projected/8969b312-9957-41a5-a640-635f79eb16e4-kube-api-access-gbztd\") pod \"neutron-db-sync-kcww7\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:51 crc kubenswrapper[4799]: I1124 08:06:51.866095 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:52 crc kubenswrapper[4799]: I1124 08:06:52.283663 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kcww7"] Nov 24 08:06:53 crc kubenswrapper[4799]: I1124 08:06:53.289347 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kcww7" event={"ID":"8969b312-9957-41a5-a640-635f79eb16e4","Type":"ContainerStarted","Data":"d572fa47053ae2af441fb1776d1b38095adb207b34655025aeeae3ca701fb086"} Nov 24 08:06:53 crc kubenswrapper[4799]: I1124 08:06:53.289614 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kcww7" event={"ID":"8969b312-9957-41a5-a640-635f79eb16e4","Type":"ContainerStarted","Data":"e493e2eff3ea41dbdc1b784d0050014bba36289f5bc2b1dcc987068dc5a237df"} Nov 24 08:06:57 crc kubenswrapper[4799]: I1124 08:06:57.328911 4799 generic.go:334] "Generic (PLEG): container finished" podID="8969b312-9957-41a5-a640-635f79eb16e4" containerID="d572fa47053ae2af441fb1776d1b38095adb207b34655025aeeae3ca701fb086" exitCode=0 Nov 24 08:06:57 crc kubenswrapper[4799]: I1124 08:06:57.328987 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kcww7" event={"ID":"8969b312-9957-41a5-a640-635f79eb16e4","Type":"ContainerDied","Data":"d572fa47053ae2af441fb1776d1b38095adb207b34655025aeeae3ca701fb086"} Nov 24 08:06:58 crc kubenswrapper[4799]: I1124 08:06:58.752782 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:58 crc kubenswrapper[4799]: I1124 08:06:58.925461 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbztd\" (UniqueName: \"kubernetes.io/projected/8969b312-9957-41a5-a640-635f79eb16e4-kube-api-access-gbztd\") pod \"8969b312-9957-41a5-a640-635f79eb16e4\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " Nov 24 08:06:58 crc kubenswrapper[4799]: I1124 08:06:58.925944 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-combined-ca-bundle\") pod \"8969b312-9957-41a5-a640-635f79eb16e4\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " Nov 24 08:06:58 crc kubenswrapper[4799]: I1124 08:06:58.925999 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-config\") pod \"8969b312-9957-41a5-a640-635f79eb16e4\" (UID: \"8969b312-9957-41a5-a640-635f79eb16e4\") " Nov 24 08:06:58 crc kubenswrapper[4799]: I1124 08:06:58.933374 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8969b312-9957-41a5-a640-635f79eb16e4-kube-api-access-gbztd" (OuterVolumeSpecName: "kube-api-access-gbztd") pod "8969b312-9957-41a5-a640-635f79eb16e4" (UID: "8969b312-9957-41a5-a640-635f79eb16e4"). InnerVolumeSpecName "kube-api-access-gbztd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:06:58 crc kubenswrapper[4799]: I1124 08:06:58.955711 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8969b312-9957-41a5-a640-635f79eb16e4" (UID: "8969b312-9957-41a5-a640-635f79eb16e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:06:58 crc kubenswrapper[4799]: I1124 08:06:58.960251 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-config" (OuterVolumeSpecName: "config") pod "8969b312-9957-41a5-a640-635f79eb16e4" (UID: "8969b312-9957-41a5-a640-635f79eb16e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.028359 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.028406 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8969b312-9957-41a5-a640-635f79eb16e4-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.028420 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbztd\" (UniqueName: \"kubernetes.io/projected/8969b312-9957-41a5-a640-635f79eb16e4-kube-api-access-gbztd\") on node \"crc\" DevicePath \"\"" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.356606 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kcww7" event={"ID":"8969b312-9957-41a5-a640-635f79eb16e4","Type":"ContainerDied","Data":"e493e2eff3ea41dbdc1b784d0050014bba36289f5bc2b1dcc987068dc5a237df"} Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.356660 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e493e2eff3ea41dbdc1b784d0050014bba36289f5bc2b1dcc987068dc5a237df" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.356762 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kcww7" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.489716 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-767b5748b7-mwgqh"] Nov 24 08:06:59 crc kubenswrapper[4799]: E1124 08:06:59.490256 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8969b312-9957-41a5-a640-635f79eb16e4" containerName="neutron-db-sync" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.490280 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8969b312-9957-41a5-a640-635f79eb16e4" containerName="neutron-db-sync" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.490523 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8969b312-9957-41a5-a640-635f79eb16e4" containerName="neutron-db-sync" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.491711 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.502521 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-767b5748b7-mwgqh"] Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.574430 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8684b7c74f-6msxl"] Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.576240 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.578607 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.578823 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.579020 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dxmcv" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.581391 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8684b7c74f-6msxl"] Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.644910 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-sb\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.644968 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-dns-svc\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.645005 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-nb\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.645056 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rpvh\" (UniqueName: \"kubernetes.io/projected/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-kube-api-access-8rpvh\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.645101 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-config\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.747168 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-nb\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.747248 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74eeb824-c788-48c1-b0d6-0d230cef156f-combined-ca-bundle\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.747290 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rpvh\" (UniqueName: \"kubernetes.io/projected/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-kube-api-access-8rpvh\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.747331 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/74eeb824-c788-48c1-b0d6-0d230cef156f-httpd-config\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.747400 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-config\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.748236 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-config\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.748251 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-nb\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.748281 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/74eeb824-c788-48c1-b0d6-0d230cef156f-config\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.748367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hfh4\" (UniqueName: \"kubernetes.io/projected/74eeb824-c788-48c1-b0d6-0d230cef156f-kube-api-access-5hfh4\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.748497 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-sb\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.748552 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-dns-svc\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.749371 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-sb\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.749667 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-dns-svc\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.768728 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rpvh\" (UniqueName: \"kubernetes.io/projected/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-kube-api-access-8rpvh\") pod \"dnsmasq-dns-767b5748b7-mwgqh\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.815283 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.850737 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/74eeb824-c788-48c1-b0d6-0d230cef156f-config\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.850800 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hfh4\" (UniqueName: \"kubernetes.io/projected/74eeb824-c788-48c1-b0d6-0d230cef156f-kube-api-access-5hfh4\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.850914 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74eeb824-c788-48c1-b0d6-0d230cef156f-combined-ca-bundle\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.850941 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/74eeb824-c788-48c1-b0d6-0d230cef156f-httpd-config\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.854814 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/74eeb824-c788-48c1-b0d6-0d230cef156f-httpd-config\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.856135 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/74eeb824-c788-48c1-b0d6-0d230cef156f-config\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.858626 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74eeb824-c788-48c1-b0d6-0d230cef156f-combined-ca-bundle\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.867885 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hfh4\" (UniqueName: \"kubernetes.io/projected/74eeb824-c788-48c1-b0d6-0d230cef156f-kube-api-access-5hfh4\") pod \"neutron-8684b7c74f-6msxl\" (UID: \"74eeb824-c788-48c1-b0d6-0d230cef156f\") " pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:06:59 crc kubenswrapper[4799]: I1124 08:06:59.896421 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:07:00 crc kubenswrapper[4799]: I1124 08:07:00.330760 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-767b5748b7-mwgqh"] Nov 24 08:07:00 crc kubenswrapper[4799]: I1124 08:07:00.369214 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" event={"ID":"71659c8f-1be8-4030-b8c4-c1ad3b3e850a","Type":"ContainerStarted","Data":"02c89c179f71174c20e5e5006059ebcc377f6b404c164a26a25a139e269cc9f2"} Nov 24 08:07:00 crc kubenswrapper[4799]: I1124 08:07:00.559708 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8684b7c74f-6msxl"] Nov 24 08:07:00 crc kubenswrapper[4799]: W1124 08:07:00.564554 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74eeb824_c788_48c1_b0d6_0d230cef156f.slice/crio-f504f79cdd273f7781221126ef1575f65f5644c02f9997c43a2c4eb37e3bfc42 WatchSource:0}: Error finding container f504f79cdd273f7781221126ef1575f65f5644c02f9997c43a2c4eb37e3bfc42: Status 404 returned error can't find the container with id f504f79cdd273f7781221126ef1575f65f5644c02f9997c43a2c4eb37e3bfc42 Nov 24 08:07:01 crc kubenswrapper[4799]: I1124 08:07:01.379143 4799 generic.go:334] "Generic (PLEG): container finished" podID="71659c8f-1be8-4030-b8c4-c1ad3b3e850a" containerID="c90bf2f1adf577afb3c7c2df06a2a01c096bb24e43a2922959e3c5de774bdf39" exitCode=0 Nov 24 08:07:01 crc kubenswrapper[4799]: I1124 08:07:01.379236 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" event={"ID":"71659c8f-1be8-4030-b8c4-c1ad3b3e850a","Type":"ContainerDied","Data":"c90bf2f1adf577afb3c7c2df06a2a01c096bb24e43a2922959e3c5de774bdf39"} Nov 24 08:07:01 crc kubenswrapper[4799]: I1124 08:07:01.381772 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8684b7c74f-6msxl" event={"ID":"74eeb824-c788-48c1-b0d6-0d230cef156f","Type":"ContainerStarted","Data":"95a908caa24bb91768e222d7da056cd2d784d1e591e40929e7d9524958c191d5"} Nov 24 08:07:01 crc kubenswrapper[4799]: I1124 08:07:01.381802 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8684b7c74f-6msxl" event={"ID":"74eeb824-c788-48c1-b0d6-0d230cef156f","Type":"ContainerStarted","Data":"f154c45b4944e5c4f0aeff0e5f669be97a8ed6f18f8109f0bf8bd3ed112752b3"} Nov 24 08:07:01 crc kubenswrapper[4799]: I1124 08:07:01.381816 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8684b7c74f-6msxl" event={"ID":"74eeb824-c788-48c1-b0d6-0d230cef156f","Type":"ContainerStarted","Data":"f504f79cdd273f7781221126ef1575f65f5644c02f9997c43a2c4eb37e3bfc42"} Nov 24 08:07:02 crc kubenswrapper[4799]: I1124 08:07:02.397755 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" event={"ID":"71659c8f-1be8-4030-b8c4-c1ad3b3e850a","Type":"ContainerStarted","Data":"09631b72971a6b9d5fed65d4c752c45991cb93f1056d637fd3b470e43ece9247"} Nov 24 08:07:02 crc kubenswrapper[4799]: I1124 08:07:02.398096 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:07:02 crc kubenswrapper[4799]: I1124 08:07:02.398117 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:07:02 crc kubenswrapper[4799]: I1124 08:07:02.421935 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8684b7c74f-6msxl" podStartSLOduration=3.421914278 podStartE2EDuration="3.421914278s" podCreationTimestamp="2025-11-24 08:06:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:07:01.446758524 +0000 UTC m=+4767.102740998" watchObservedRunningTime="2025-11-24 08:07:02.421914278 +0000 UTC m=+4768.077896772" Nov 24 08:07:02 crc kubenswrapper[4799]: I1124 08:07:02.426392 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" podStartSLOduration=3.426373392 podStartE2EDuration="3.426373392s" podCreationTimestamp="2025-11-24 08:06:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:07:02.418187835 +0000 UTC m=+4768.074170319" watchObservedRunningTime="2025-11-24 08:07:02.426373392 +0000 UTC m=+4768.082355866" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.325668 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dfpdb"] Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.328992 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.336241 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dfpdb"] Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.419320 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9xt4\" (UniqueName: \"kubernetes.io/projected/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-kube-api-access-g9xt4\") pod \"redhat-operators-dfpdb\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.419360 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-utilities\") pod \"redhat-operators-dfpdb\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.419392 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-catalog-content\") pod \"redhat-operators-dfpdb\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.520983 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9xt4\" (UniqueName: \"kubernetes.io/projected/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-kube-api-access-g9xt4\") pod \"redhat-operators-dfpdb\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.521029 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-utilities\") pod \"redhat-operators-dfpdb\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.521058 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-catalog-content\") pod \"redhat-operators-dfpdb\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.521565 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-catalog-content\") pod \"redhat-operators-dfpdb\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.521742 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-utilities\") pod \"redhat-operators-dfpdb\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.550170 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9xt4\" (UniqueName: \"kubernetes.io/projected/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-kube-api-access-g9xt4\") pod \"redhat-operators-dfpdb\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.670153 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.819011 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.872643 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d59b85bf-kfkc9"] Nov 24 08:07:09 crc kubenswrapper[4799]: I1124 08:07:09.873176 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" podUID="375fe805-7346-48ac-a330-f28ff5da9303" containerName="dnsmasq-dns" containerID="cri-o://065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4" gracePeriod=10 Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.181070 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dfpdb"] Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.338524 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.436674 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-dns-svc\") pod \"375fe805-7346-48ac-a330-f28ff5da9303\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.437022 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blx8j\" (UniqueName: \"kubernetes.io/projected/375fe805-7346-48ac-a330-f28ff5da9303-kube-api-access-blx8j\") pod \"375fe805-7346-48ac-a330-f28ff5da9303\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.437592 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-sb\") pod \"375fe805-7346-48ac-a330-f28ff5da9303\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.437729 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-config\") pod \"375fe805-7346-48ac-a330-f28ff5da9303\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.437807 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-nb\") pod \"375fe805-7346-48ac-a330-f28ff5da9303\" (UID: \"375fe805-7346-48ac-a330-f28ff5da9303\") " Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.444296 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/375fe805-7346-48ac-a330-f28ff5da9303-kube-api-access-blx8j" (OuterVolumeSpecName: "kube-api-access-blx8j") pod "375fe805-7346-48ac-a330-f28ff5da9303" (UID: "375fe805-7346-48ac-a330-f28ff5da9303"). InnerVolumeSpecName "kube-api-access-blx8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.479839 4799 generic.go:334] "Generic (PLEG): container finished" podID="375fe805-7346-48ac-a330-f28ff5da9303" containerID="065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4" exitCode=0 Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.479949 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" event={"ID":"375fe805-7346-48ac-a330-f28ff5da9303","Type":"ContainerDied","Data":"065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4"} Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.479983 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" event={"ID":"375fe805-7346-48ac-a330-f28ff5da9303","Type":"ContainerDied","Data":"b82fdc447f782bd137e330ec34d98b9209aaf5a6f36cd6372346829468f407d9"} Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.480003 4799 scope.go:117] "RemoveContainer" containerID="065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.480153 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d59b85bf-kfkc9" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.488598 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfpdb" event={"ID":"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f","Type":"ContainerStarted","Data":"aec5c7ab6909c3e59bb4590bd4bce3b8935654600943c59a67ab1070a06a5795"} Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.488657 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfpdb" event={"ID":"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f","Type":"ContainerStarted","Data":"a1bc77db4f338e781628c61289627db9d96d4db9cb095381dab2b8a80ea0ae54"} Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.489674 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "375fe805-7346-48ac-a330-f28ff5da9303" (UID: "375fe805-7346-48ac-a330-f28ff5da9303"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.491351 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "375fe805-7346-48ac-a330-f28ff5da9303" (UID: "375fe805-7346-48ac-a330-f28ff5da9303"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.495974 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "375fe805-7346-48ac-a330-f28ff5da9303" (UID: "375fe805-7346-48ac-a330-f28ff5da9303"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.512448 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-config" (OuterVolumeSpecName: "config") pod "375fe805-7346-48ac-a330-f28ff5da9303" (UID: "375fe805-7346-48ac-a330-f28ff5da9303"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.529606 4799 scope.go:117] "RemoveContainer" containerID="0c92f88b54076937c229edca66bebf876ca670b468c18ca083e68a776ef22cfc" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.539767 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.539813 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.539828 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.539839 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/375fe805-7346-48ac-a330-f28ff5da9303-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.539872 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blx8j\" (UniqueName: \"kubernetes.io/projected/375fe805-7346-48ac-a330-f28ff5da9303-kube-api-access-blx8j\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.555953 4799 scope.go:117] "RemoveContainer" containerID="065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4" Nov 24 08:07:10 crc kubenswrapper[4799]: E1124 08:07:10.556462 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4\": container with ID starting with 065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4 not found: ID does not exist" containerID="065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.556600 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4"} err="failed to get container status \"065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4\": rpc error: code = NotFound desc = could not find container \"065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4\": container with ID starting with 065986da2b95273c6f757b66196b1cdf8bff42477b57144e1b3410fb8c2e77d4 not found: ID does not exist" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.556703 4799 scope.go:117] "RemoveContainer" containerID="0c92f88b54076937c229edca66bebf876ca670b468c18ca083e68a776ef22cfc" Nov 24 08:07:10 crc kubenswrapper[4799]: E1124 08:07:10.556989 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c92f88b54076937c229edca66bebf876ca670b468c18ca083e68a776ef22cfc\": container with ID starting with 0c92f88b54076937c229edca66bebf876ca670b468c18ca083e68a776ef22cfc not found: ID does not exist" containerID="0c92f88b54076937c229edca66bebf876ca670b468c18ca083e68a776ef22cfc" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.557115 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c92f88b54076937c229edca66bebf876ca670b468c18ca083e68a776ef22cfc"} err="failed to get container status \"0c92f88b54076937c229edca66bebf876ca670b468c18ca083e68a776ef22cfc\": rpc error: code = NotFound desc = could not find container \"0c92f88b54076937c229edca66bebf876ca670b468c18ca083e68a776ef22cfc\": container with ID starting with 0c92f88b54076937c229edca66bebf876ca670b468c18ca083e68a776ef22cfc not found: ID does not exist" Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.837403 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d59b85bf-kfkc9"] Nov 24 08:07:10 crc kubenswrapper[4799]: I1124 08:07:10.854339 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68d59b85bf-kfkc9"] Nov 24 08:07:11 crc kubenswrapper[4799]: I1124 08:07:11.501105 4799 generic.go:334] "Generic (PLEG): container finished" podID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" containerID="aec5c7ab6909c3e59bb4590bd4bce3b8935654600943c59a67ab1070a06a5795" exitCode=0 Nov 24 08:07:11 crc kubenswrapper[4799]: I1124 08:07:11.501159 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfpdb" event={"ID":"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f","Type":"ContainerDied","Data":"aec5c7ab6909c3e59bb4590bd4bce3b8935654600943c59a67ab1070a06a5795"} Nov 24 08:07:11 crc kubenswrapper[4799]: I1124 08:07:11.501189 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfpdb" event={"ID":"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f","Type":"ContainerStarted","Data":"137e596f02ac55548afca875db00cd37bcef268de0783665a362bd55ee683466"} Nov 24 08:07:11 crc kubenswrapper[4799]: I1124 08:07:11.639061 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="375fe805-7346-48ac-a330-f28ff5da9303" path="/var/lib/kubelet/pods/375fe805-7346-48ac-a330-f28ff5da9303/volumes" Nov 24 08:07:12 crc kubenswrapper[4799]: I1124 08:07:12.510264 4799 generic.go:334] "Generic (PLEG): container finished" podID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" containerID="137e596f02ac55548afca875db00cd37bcef268de0783665a362bd55ee683466" exitCode=0 Nov 24 08:07:12 crc kubenswrapper[4799]: I1124 08:07:12.510381 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfpdb" event={"ID":"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f","Type":"ContainerDied","Data":"137e596f02ac55548afca875db00cd37bcef268de0783665a362bd55ee683466"} Nov 24 08:07:13 crc kubenswrapper[4799]: I1124 08:07:13.523463 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfpdb" event={"ID":"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f","Type":"ContainerStarted","Data":"7f1f8b0d04498e505a2b91e7c45d66809bda6d5fc003fa06a0253c9adea934a3"} Nov 24 08:07:13 crc kubenswrapper[4799]: I1124 08:07:13.542669 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dfpdb" podStartSLOduration=1.967966826 podStartE2EDuration="4.542633902s" podCreationTimestamp="2025-11-24 08:07:09 +0000 UTC" firstStartedPulling="2025-11-24 08:07:10.493099222 +0000 UTC m=+4776.149081696" lastFinishedPulling="2025-11-24 08:07:13.067766288 +0000 UTC m=+4778.723748772" observedRunningTime="2025-11-24 08:07:13.541570553 +0000 UTC m=+4779.197553047" watchObservedRunningTime="2025-11-24 08:07:13.542633902 +0000 UTC m=+4779.198616376" Nov 24 08:07:19 crc kubenswrapper[4799]: I1124 08:07:19.671331 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:19 crc kubenswrapper[4799]: I1124 08:07:19.671856 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:19 crc kubenswrapper[4799]: I1124 08:07:19.717647 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:20 crc kubenswrapper[4799]: I1124 08:07:20.400996 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:07:20 crc kubenswrapper[4799]: I1124 08:07:20.401061 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:07:20 crc kubenswrapper[4799]: I1124 08:07:20.401105 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 08:07:20 crc kubenswrapper[4799]: I1124 08:07:20.402005 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"023b79a79ebb461005d32f47310810404d347caf336f616e7c3a30dfcc69e291"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:07:20 crc kubenswrapper[4799]: I1124 08:07:20.402086 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://023b79a79ebb461005d32f47310810404d347caf336f616e7c3a30dfcc69e291" gracePeriod=600 Nov 24 08:07:20 crc kubenswrapper[4799]: I1124 08:07:20.585230 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="023b79a79ebb461005d32f47310810404d347caf336f616e7c3a30dfcc69e291" exitCode=0 Nov 24 08:07:20 crc kubenswrapper[4799]: I1124 08:07:20.585362 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"023b79a79ebb461005d32f47310810404d347caf336f616e7c3a30dfcc69e291"} Nov 24 08:07:20 crc kubenswrapper[4799]: I1124 08:07:20.585716 4799 scope.go:117] "RemoveContainer" containerID="ec356352bc5062bcb6c7e807ce857ef7f2ea0b384c38b2dace5c9e579d634323" Nov 24 08:07:20 crc kubenswrapper[4799]: I1124 08:07:20.635140 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:20 crc kubenswrapper[4799]: I1124 08:07:20.678703 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dfpdb"] Nov 24 08:07:21 crc kubenswrapper[4799]: I1124 08:07:21.599982 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329"} Nov 24 08:07:22 crc kubenswrapper[4799]: I1124 08:07:22.608260 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dfpdb" podUID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" containerName="registry-server" containerID="cri-o://7f1f8b0d04498e505a2b91e7c45d66809bda6d5fc003fa06a0253c9adea934a3" gracePeriod=2 Nov 24 08:07:23 crc kubenswrapper[4799]: I1124 08:07:23.617251 4799 generic.go:334] "Generic (PLEG): container finished" podID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" containerID="7f1f8b0d04498e505a2b91e7c45d66809bda6d5fc003fa06a0253c9adea934a3" exitCode=0 Nov 24 08:07:23 crc kubenswrapper[4799]: I1124 08:07:23.617320 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfpdb" event={"ID":"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f","Type":"ContainerDied","Data":"7f1f8b0d04498e505a2b91e7c45d66809bda6d5fc003fa06a0253c9adea934a3"} Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.153802 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.282443 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9xt4\" (UniqueName: \"kubernetes.io/projected/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-kube-api-access-g9xt4\") pod \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.282496 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-utilities\") pod \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.282550 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-catalog-content\") pod \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\" (UID: \"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f\") " Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.283750 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-utilities" (OuterVolumeSpecName: "utilities") pod "f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" (UID: "f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.289520 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-kube-api-access-g9xt4" (OuterVolumeSpecName: "kube-api-access-g9xt4") pod "f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" (UID: "f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f"). InnerVolumeSpecName "kube-api-access-g9xt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.384365 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9xt4\" (UniqueName: \"kubernetes.io/projected/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-kube-api-access-g9xt4\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.384400 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.473810 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" (UID: "f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.485880 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.629393 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfpdb" event={"ID":"f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f","Type":"ContainerDied","Data":"a1bc77db4f338e781628c61289627db9d96d4db9cb095381dab2b8a80ea0ae54"} Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.629439 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfpdb" Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.629448 4799 scope.go:117] "RemoveContainer" containerID="7f1f8b0d04498e505a2b91e7c45d66809bda6d5fc003fa06a0253c9adea934a3" Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.654567 4799 scope.go:117] "RemoveContainer" containerID="137e596f02ac55548afca875db00cd37bcef268de0783665a362bd55ee683466" Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.666127 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dfpdb"] Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.676192 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dfpdb"] Nov 24 08:07:24 crc kubenswrapper[4799]: I1124 08:07:24.701792 4799 scope.go:117] "RemoveContainer" containerID="aec5c7ab6909c3e59bb4590bd4bce3b8935654600943c59a67ab1070a06a5795" Nov 24 08:07:25 crc kubenswrapper[4799]: I1124 08:07:25.645921 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" path="/var/lib/kubelet/pods/f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f/volumes" Nov 24 08:07:29 crc kubenswrapper[4799]: I1124 08:07:29.913758 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8684b7c74f-6msxl" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.532660 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-lcv54"] Nov 24 08:07:37 crc kubenswrapper[4799]: E1124 08:07:37.533486 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" containerName="extract-content" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.533499 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" containerName="extract-content" Nov 24 08:07:37 crc kubenswrapper[4799]: E1124 08:07:37.533512 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" containerName="registry-server" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.533518 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" containerName="registry-server" Nov 24 08:07:37 crc kubenswrapper[4799]: E1124 08:07:37.533529 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="375fe805-7346-48ac-a330-f28ff5da9303" containerName="init" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.533535 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="375fe805-7346-48ac-a330-f28ff5da9303" containerName="init" Nov 24 08:07:37 crc kubenswrapper[4799]: E1124 08:07:37.533546 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" containerName="extract-utilities" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.533552 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" containerName="extract-utilities" Nov 24 08:07:37 crc kubenswrapper[4799]: E1124 08:07:37.533568 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="375fe805-7346-48ac-a330-f28ff5da9303" containerName="dnsmasq-dns" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.533573 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="375fe805-7346-48ac-a330-f28ff5da9303" containerName="dnsmasq-dns" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.533729 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="375fe805-7346-48ac-a330-f28ff5da9303" containerName="dnsmasq-dns" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.533742 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f90aeb12-e85a-4a3f-81ef-d28aa05d1c1f" containerName="registry-server" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.534274 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lcv54" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.556052 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-lcv54"] Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.629666 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx5db\" (UniqueName: \"kubernetes.io/projected/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-kube-api-access-xx5db\") pod \"glance-db-create-lcv54\" (UID: \"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35\") " pod="openstack/glance-db-create-lcv54" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.629709 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-operator-scripts\") pod \"glance-db-create-lcv54\" (UID: \"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35\") " pod="openstack/glance-db-create-lcv54" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.640119 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-64b3-account-create-vrs9s"] Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.641493 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-64b3-account-create-vrs9s" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.644488 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.645736 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-64b3-account-create-vrs9s"] Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.731423 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksl8v\" (UniqueName: \"kubernetes.io/projected/1613cebb-f738-4ad3-bdcb-9d9553c0505a-kube-api-access-ksl8v\") pod \"glance-64b3-account-create-vrs9s\" (UID: \"1613cebb-f738-4ad3-bdcb-9d9553c0505a\") " pod="openstack/glance-64b3-account-create-vrs9s" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.731476 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx5db\" (UniqueName: \"kubernetes.io/projected/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-kube-api-access-xx5db\") pod \"glance-db-create-lcv54\" (UID: \"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35\") " pod="openstack/glance-db-create-lcv54" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.731508 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-operator-scripts\") pod \"glance-db-create-lcv54\" (UID: \"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35\") " pod="openstack/glance-db-create-lcv54" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.731580 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1613cebb-f738-4ad3-bdcb-9d9553c0505a-operator-scripts\") pod \"glance-64b3-account-create-vrs9s\" (UID: \"1613cebb-f738-4ad3-bdcb-9d9553c0505a\") " pod="openstack/glance-64b3-account-create-vrs9s" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.732355 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-operator-scripts\") pod \"glance-db-create-lcv54\" (UID: \"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35\") " pod="openstack/glance-db-create-lcv54" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.762127 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx5db\" (UniqueName: \"kubernetes.io/projected/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-kube-api-access-xx5db\") pod \"glance-db-create-lcv54\" (UID: \"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35\") " pod="openstack/glance-db-create-lcv54" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.833696 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksl8v\" (UniqueName: \"kubernetes.io/projected/1613cebb-f738-4ad3-bdcb-9d9553c0505a-kube-api-access-ksl8v\") pod \"glance-64b3-account-create-vrs9s\" (UID: \"1613cebb-f738-4ad3-bdcb-9d9553c0505a\") " pod="openstack/glance-64b3-account-create-vrs9s" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.833862 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1613cebb-f738-4ad3-bdcb-9d9553c0505a-operator-scripts\") pod \"glance-64b3-account-create-vrs9s\" (UID: \"1613cebb-f738-4ad3-bdcb-9d9553c0505a\") " pod="openstack/glance-64b3-account-create-vrs9s" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.834782 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1613cebb-f738-4ad3-bdcb-9d9553c0505a-operator-scripts\") pod \"glance-64b3-account-create-vrs9s\" (UID: \"1613cebb-f738-4ad3-bdcb-9d9553c0505a\") " pod="openstack/glance-64b3-account-create-vrs9s" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.861653 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksl8v\" (UniqueName: \"kubernetes.io/projected/1613cebb-f738-4ad3-bdcb-9d9553c0505a-kube-api-access-ksl8v\") pod \"glance-64b3-account-create-vrs9s\" (UID: \"1613cebb-f738-4ad3-bdcb-9d9553c0505a\") " pod="openstack/glance-64b3-account-create-vrs9s" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.875712 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lcv54" Nov 24 08:07:37 crc kubenswrapper[4799]: I1124 08:07:37.971568 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-64b3-account-create-vrs9s" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.331747 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-lcv54"] Nov 24 08:07:38 crc kubenswrapper[4799]: W1124 08:07:38.337992 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3bea8cc5_3dd3_4719_8b6e_cdb66baa4a35.slice/crio-3f66a61f1eba35774947d10b4015a169e26098b21985025b25f91a40abca880d WatchSource:0}: Error finding container 3f66a61f1eba35774947d10b4015a169e26098b21985025b25f91a40abca880d: Status 404 returned error can't find the container with id 3f66a61f1eba35774947d10b4015a169e26098b21985025b25f91a40abca880d Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.468192 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sntrt"] Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.471081 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.479088 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-64b3-account-create-vrs9s"] Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.492033 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sntrt"] Nov 24 08:07:38 crc kubenswrapper[4799]: W1124 08:07:38.497740 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1613cebb_f738_4ad3_bdcb_9d9553c0505a.slice/crio-7edf85e4c67d182548146aff016f13db4436ccc1ab4cb3aead3cea56089318d5 WatchSource:0}: Error finding container 7edf85e4c67d182548146aff016f13db4436ccc1ab4cb3aead3cea56089318d5: Status 404 returned error can't find the container with id 7edf85e4c67d182548146aff016f13db4436ccc1ab4cb3aead3cea56089318d5 Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.552228 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z6qd\" (UniqueName: \"kubernetes.io/projected/22864cc4-2d09-49e6-b55c-d10295e62c8e-kube-api-access-5z6qd\") pod \"certified-operators-sntrt\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.553271 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-catalog-content\") pod \"certified-operators-sntrt\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.553402 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-utilities\") pod \"certified-operators-sntrt\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.655074 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-catalog-content\") pod \"certified-operators-sntrt\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.655329 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-utilities\") pod \"certified-operators-sntrt\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.655491 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z6qd\" (UniqueName: \"kubernetes.io/projected/22864cc4-2d09-49e6-b55c-d10295e62c8e-kube-api-access-5z6qd\") pod \"certified-operators-sntrt\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.656251 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-catalog-content\") pod \"certified-operators-sntrt\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.656546 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-utilities\") pod \"certified-operators-sntrt\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.677478 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z6qd\" (UniqueName: \"kubernetes.io/projected/22864cc4-2d09-49e6-b55c-d10295e62c8e-kube-api-access-5z6qd\") pod \"certified-operators-sntrt\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.760717 4799 generic.go:334] "Generic (PLEG): container finished" podID="3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35" containerID="1dc3fc380db55fd127a6133a8928eb1cdcc9be66e2dbe95072bd3574e2cc5235" exitCode=0 Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.760790 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lcv54" event={"ID":"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35","Type":"ContainerDied","Data":"1dc3fc380db55fd127a6133a8928eb1cdcc9be66e2dbe95072bd3574e2cc5235"} Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.760817 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lcv54" event={"ID":"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35","Type":"ContainerStarted","Data":"3f66a61f1eba35774947d10b4015a169e26098b21985025b25f91a40abca880d"} Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.762238 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-64b3-account-create-vrs9s" event={"ID":"1613cebb-f738-4ad3-bdcb-9d9553c0505a","Type":"ContainerStarted","Data":"d7e327782ee7420167fc4fa23e5b79e12160ed533afde1fe1838da935cc9f6cb"} Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.762286 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-64b3-account-create-vrs9s" event={"ID":"1613cebb-f738-4ad3-bdcb-9d9553c0505a","Type":"ContainerStarted","Data":"7edf85e4c67d182548146aff016f13db4436ccc1ab4cb3aead3cea56089318d5"} Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.793272 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:38 crc kubenswrapper[4799]: I1124 08:07:38.799424 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-64b3-account-create-vrs9s" podStartSLOduration=1.799402578 podStartE2EDuration="1.799402578s" podCreationTimestamp="2025-11-24 08:07:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:07:38.790229493 +0000 UTC m=+4804.446211957" watchObservedRunningTime="2025-11-24 08:07:38.799402578 +0000 UTC m=+4804.455385052" Nov 24 08:07:39 crc kubenswrapper[4799]: I1124 08:07:39.245503 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sntrt"] Nov 24 08:07:39 crc kubenswrapper[4799]: W1124 08:07:39.252292 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22864cc4_2d09_49e6_b55c_d10295e62c8e.slice/crio-f4734faf913a5193e8bd0735133ea61036ba1ccef98727dd353f64fbae1bcccf WatchSource:0}: Error finding container f4734faf913a5193e8bd0735133ea61036ba1ccef98727dd353f64fbae1bcccf: Status 404 returned error can't find the container with id f4734faf913a5193e8bd0735133ea61036ba1ccef98727dd353f64fbae1bcccf Nov 24 08:07:39 crc kubenswrapper[4799]: I1124 08:07:39.776356 4799 generic.go:334] "Generic (PLEG): container finished" podID="1613cebb-f738-4ad3-bdcb-9d9553c0505a" containerID="d7e327782ee7420167fc4fa23e5b79e12160ed533afde1fe1838da935cc9f6cb" exitCode=0 Nov 24 08:07:39 crc kubenswrapper[4799]: I1124 08:07:39.776486 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-64b3-account-create-vrs9s" event={"ID":"1613cebb-f738-4ad3-bdcb-9d9553c0505a","Type":"ContainerDied","Data":"d7e327782ee7420167fc4fa23e5b79e12160ed533afde1fe1838da935cc9f6cb"} Nov 24 08:07:39 crc kubenswrapper[4799]: I1124 08:07:39.780340 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sntrt" event={"ID":"22864cc4-2d09-49e6-b55c-d10295e62c8e","Type":"ContainerStarted","Data":"26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3"} Nov 24 08:07:39 crc kubenswrapper[4799]: I1124 08:07:39.780407 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sntrt" event={"ID":"22864cc4-2d09-49e6-b55c-d10295e62c8e","Type":"ContainerStarted","Data":"f4734faf913a5193e8bd0735133ea61036ba1ccef98727dd353f64fbae1bcccf"} Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.072152 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lcv54" Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.183169 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx5db\" (UniqueName: \"kubernetes.io/projected/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-kube-api-access-xx5db\") pod \"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35\" (UID: \"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35\") " Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.183227 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-operator-scripts\") pod \"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35\" (UID: \"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35\") " Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.185536 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35" (UID: "3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.193206 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-kube-api-access-xx5db" (OuterVolumeSpecName: "kube-api-access-xx5db") pod "3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35" (UID: "3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35"). InnerVolumeSpecName "kube-api-access-xx5db". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.285450 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.285490 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx5db\" (UniqueName: \"kubernetes.io/projected/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35-kube-api-access-xx5db\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.794222 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lcv54" event={"ID":"3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35","Type":"ContainerDied","Data":"3f66a61f1eba35774947d10b4015a169e26098b21985025b25f91a40abca880d"} Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.794824 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f66a61f1eba35774947d10b4015a169e26098b21985025b25f91a40abca880d" Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.794291 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lcv54" Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.797309 4799 generic.go:334] "Generic (PLEG): container finished" podID="22864cc4-2d09-49e6-b55c-d10295e62c8e" containerID="26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3" exitCode=0 Nov 24 08:07:40 crc kubenswrapper[4799]: I1124 08:07:40.797490 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sntrt" event={"ID":"22864cc4-2d09-49e6-b55c-d10295e62c8e","Type":"ContainerDied","Data":"26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3"} Nov 24 08:07:41 crc kubenswrapper[4799]: I1124 08:07:41.128945 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-64b3-account-create-vrs9s" Nov 24 08:07:41 crc kubenswrapper[4799]: I1124 08:07:41.202062 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1613cebb-f738-4ad3-bdcb-9d9553c0505a-operator-scripts\") pod \"1613cebb-f738-4ad3-bdcb-9d9553c0505a\" (UID: \"1613cebb-f738-4ad3-bdcb-9d9553c0505a\") " Nov 24 08:07:41 crc kubenswrapper[4799]: I1124 08:07:41.202124 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksl8v\" (UniqueName: \"kubernetes.io/projected/1613cebb-f738-4ad3-bdcb-9d9553c0505a-kube-api-access-ksl8v\") pod \"1613cebb-f738-4ad3-bdcb-9d9553c0505a\" (UID: \"1613cebb-f738-4ad3-bdcb-9d9553c0505a\") " Nov 24 08:07:41 crc kubenswrapper[4799]: I1124 08:07:41.203159 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1613cebb-f738-4ad3-bdcb-9d9553c0505a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1613cebb-f738-4ad3-bdcb-9d9553c0505a" (UID: "1613cebb-f738-4ad3-bdcb-9d9553c0505a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:07:41 crc kubenswrapper[4799]: I1124 08:07:41.208044 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1613cebb-f738-4ad3-bdcb-9d9553c0505a-kube-api-access-ksl8v" (OuterVolumeSpecName: "kube-api-access-ksl8v") pod "1613cebb-f738-4ad3-bdcb-9d9553c0505a" (UID: "1613cebb-f738-4ad3-bdcb-9d9553c0505a"). InnerVolumeSpecName "kube-api-access-ksl8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:07:41 crc kubenswrapper[4799]: I1124 08:07:41.304399 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1613cebb-f738-4ad3-bdcb-9d9553c0505a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:41 crc kubenswrapper[4799]: I1124 08:07:41.304479 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksl8v\" (UniqueName: \"kubernetes.io/projected/1613cebb-f738-4ad3-bdcb-9d9553c0505a-kube-api-access-ksl8v\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:41 crc kubenswrapper[4799]: I1124 08:07:41.817198 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-64b3-account-create-vrs9s" Nov 24 08:07:41 crc kubenswrapper[4799]: I1124 08:07:41.817220 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-64b3-account-create-vrs9s" event={"ID":"1613cebb-f738-4ad3-bdcb-9d9553c0505a","Type":"ContainerDied","Data":"7edf85e4c67d182548146aff016f13db4436ccc1ab4cb3aead3cea56089318d5"} Nov 24 08:07:41 crc kubenswrapper[4799]: I1124 08:07:41.818001 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7edf85e4c67d182548146aff016f13db4436ccc1ab4cb3aead3cea56089318d5" Nov 24 08:07:41 crc kubenswrapper[4799]: I1124 08:07:41.825937 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sntrt" event={"ID":"22864cc4-2d09-49e6-b55c-d10295e62c8e","Type":"ContainerStarted","Data":"b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618"} Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.802816 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-jl8c5"] Nov 24 08:07:42 crc kubenswrapper[4799]: E1124 08:07:42.803635 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1613cebb-f738-4ad3-bdcb-9d9553c0505a" containerName="mariadb-account-create" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.803681 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1613cebb-f738-4ad3-bdcb-9d9553c0505a" containerName="mariadb-account-create" Nov 24 08:07:42 crc kubenswrapper[4799]: E1124 08:07:42.803721 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35" containerName="mariadb-database-create" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.803734 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35" containerName="mariadb-database-create" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.804093 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1613cebb-f738-4ad3-bdcb-9d9553c0505a" containerName="mariadb-account-create" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.804139 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35" containerName="mariadb-database-create" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.805193 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.807830 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lwt7f" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.809369 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.826133 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jl8c5"] Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.835630 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-db-sync-config-data\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.835700 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgngg\" (UniqueName: \"kubernetes.io/projected/8f99a056-2251-45ca-a55e-f970ab0fe1e2-kube-api-access-zgngg\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.835777 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-combined-ca-bundle\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.835801 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-config-data\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.848828 4799 generic.go:334] "Generic (PLEG): container finished" podID="22864cc4-2d09-49e6-b55c-d10295e62c8e" containerID="b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618" exitCode=0 Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.848888 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sntrt" event={"ID":"22864cc4-2d09-49e6-b55c-d10295e62c8e","Type":"ContainerDied","Data":"b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618"} Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.936504 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-db-sync-config-data\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.936862 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgngg\" (UniqueName: \"kubernetes.io/projected/8f99a056-2251-45ca-a55e-f970ab0fe1e2-kube-api-access-zgngg\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.936924 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-combined-ca-bundle\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.936946 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-config-data\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.942550 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-db-sync-config-data\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.943219 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-config-data\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.946554 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-combined-ca-bundle\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:42 crc kubenswrapper[4799]: I1124 08:07:42.959493 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgngg\" (UniqueName: \"kubernetes.io/projected/8f99a056-2251-45ca-a55e-f970ab0fe1e2-kube-api-access-zgngg\") pod \"glance-db-sync-jl8c5\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:43 crc kubenswrapper[4799]: I1124 08:07:43.133937 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jl8c5" Nov 24 08:07:43 crc kubenswrapper[4799]: I1124 08:07:43.858606 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sntrt" event={"ID":"22864cc4-2d09-49e6-b55c-d10295e62c8e","Type":"ContainerStarted","Data":"7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d"} Nov 24 08:07:43 crc kubenswrapper[4799]: I1124 08:07:43.887473 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sntrt" podStartSLOduration=3.391616391 podStartE2EDuration="5.887453277s" podCreationTimestamp="2025-11-24 08:07:38 +0000 UTC" firstStartedPulling="2025-11-24 08:07:40.802198175 +0000 UTC m=+4806.458180659" lastFinishedPulling="2025-11-24 08:07:43.298035071 +0000 UTC m=+4808.954017545" observedRunningTime="2025-11-24 08:07:43.880338129 +0000 UTC m=+4809.536320603" watchObservedRunningTime="2025-11-24 08:07:43.887453277 +0000 UTC m=+4809.543435751" Nov 24 08:07:44 crc kubenswrapper[4799]: I1124 08:07:44.118348 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jl8c5"] Nov 24 08:07:44 crc kubenswrapper[4799]: I1124 08:07:44.873285 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jl8c5" event={"ID":"8f99a056-2251-45ca-a55e-f970ab0fe1e2","Type":"ContainerStarted","Data":"741757f8935f0314af3ba153f1d0fbfbd255244bd5b4354963e49b52eba1f0fc"} Nov 24 08:07:48 crc kubenswrapper[4799]: I1124 08:07:48.794152 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:48 crc kubenswrapper[4799]: I1124 08:07:48.795101 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:48 crc kubenswrapper[4799]: I1124 08:07:48.847569 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:48 crc kubenswrapper[4799]: I1124 08:07:48.948015 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:49 crc kubenswrapper[4799]: I1124 08:07:49.081353 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sntrt"] Nov 24 08:07:50 crc kubenswrapper[4799]: I1124 08:07:50.918335 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sntrt" podUID="22864cc4-2d09-49e6-b55c-d10295e62c8e" containerName="registry-server" containerID="cri-o://7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d" gracePeriod=2 Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.436461 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.499401 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-catalog-content\") pod \"22864cc4-2d09-49e6-b55c-d10295e62c8e\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.499509 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-utilities\") pod \"22864cc4-2d09-49e6-b55c-d10295e62c8e\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.499563 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z6qd\" (UniqueName: \"kubernetes.io/projected/22864cc4-2d09-49e6-b55c-d10295e62c8e-kube-api-access-5z6qd\") pod \"22864cc4-2d09-49e6-b55c-d10295e62c8e\" (UID: \"22864cc4-2d09-49e6-b55c-d10295e62c8e\") " Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.500966 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-utilities" (OuterVolumeSpecName: "utilities") pod "22864cc4-2d09-49e6-b55c-d10295e62c8e" (UID: "22864cc4-2d09-49e6-b55c-d10295e62c8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.506050 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22864cc4-2d09-49e6-b55c-d10295e62c8e-kube-api-access-5z6qd" (OuterVolumeSpecName: "kube-api-access-5z6qd") pod "22864cc4-2d09-49e6-b55c-d10295e62c8e" (UID: "22864cc4-2d09-49e6-b55c-d10295e62c8e"). InnerVolumeSpecName "kube-api-access-5z6qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.563865 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22864cc4-2d09-49e6-b55c-d10295e62c8e" (UID: "22864cc4-2d09-49e6-b55c-d10295e62c8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.601747 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.601788 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22864cc4-2d09-49e6-b55c-d10295e62c8e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.601799 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z6qd\" (UniqueName: \"kubernetes.io/projected/22864cc4-2d09-49e6-b55c-d10295e62c8e-kube-api-access-5z6qd\") on node \"crc\" DevicePath \"\"" Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.928572 4799 generic.go:334] "Generic (PLEG): container finished" podID="22864cc4-2d09-49e6-b55c-d10295e62c8e" containerID="7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d" exitCode=0 Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.928722 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sntrt" event={"ID":"22864cc4-2d09-49e6-b55c-d10295e62c8e","Type":"ContainerDied","Data":"7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d"} Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.929599 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sntrt" event={"ID":"22864cc4-2d09-49e6-b55c-d10295e62c8e","Type":"ContainerDied","Data":"f4734faf913a5193e8bd0735133ea61036ba1ccef98727dd353f64fbae1bcccf"} Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.929626 4799 scope.go:117] "RemoveContainer" containerID="7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d" Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.928787 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sntrt" Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.959721 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sntrt"] Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.960807 4799 scope.go:117] "RemoveContainer" containerID="b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618" Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.967549 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sntrt"] Nov 24 08:07:51 crc kubenswrapper[4799]: I1124 08:07:51.978808 4799 scope.go:117] "RemoveContainer" containerID="26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3" Nov 24 08:07:52 crc kubenswrapper[4799]: I1124 08:07:52.019236 4799 scope.go:117] "RemoveContainer" containerID="7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d" Nov 24 08:07:52 crc kubenswrapper[4799]: E1124 08:07:52.019641 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d\": container with ID starting with 7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d not found: ID does not exist" containerID="7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d" Nov 24 08:07:52 crc kubenswrapper[4799]: I1124 08:07:52.019686 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d"} err="failed to get container status \"7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d\": rpc error: code = NotFound desc = could not find container \"7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d\": container with ID starting with 7b0c1cc19b234ac66600fc171906ec72c7beca3e61676f5f845f641244450a9d not found: ID does not exist" Nov 24 08:07:52 crc kubenswrapper[4799]: I1124 08:07:52.019716 4799 scope.go:117] "RemoveContainer" containerID="b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618" Nov 24 08:07:52 crc kubenswrapper[4799]: E1124 08:07:52.020065 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618\": container with ID starting with b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618 not found: ID does not exist" containerID="b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618" Nov 24 08:07:52 crc kubenswrapper[4799]: I1124 08:07:52.020094 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618"} err="failed to get container status \"b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618\": rpc error: code = NotFound desc = could not find container \"b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618\": container with ID starting with b1f0fb43aaf197f81792e12a27006415dc0079b82c1bca5eef2a5a0f579f9618 not found: ID does not exist" Nov 24 08:07:52 crc kubenswrapper[4799]: I1124 08:07:52.020114 4799 scope.go:117] "RemoveContainer" containerID="26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3" Nov 24 08:07:52 crc kubenswrapper[4799]: E1124 08:07:52.020409 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3\": container with ID starting with 26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3 not found: ID does not exist" containerID="26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3" Nov 24 08:07:52 crc kubenswrapper[4799]: I1124 08:07:52.020446 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3"} err="failed to get container status \"26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3\": rpc error: code = NotFound desc = could not find container \"26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3\": container with ID starting with 26072cba7acfffc121c3594baf80bdbc5942f3f44f9ff597bbfcc9b523bf66d3 not found: ID does not exist" Nov 24 08:07:53 crc kubenswrapper[4799]: I1124 08:07:53.637802 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22864cc4-2d09-49e6-b55c-d10295e62c8e" path="/var/lib/kubelet/pods/22864cc4-2d09-49e6-b55c-d10295e62c8e/volumes" Nov 24 08:08:01 crc kubenswrapper[4799]: I1124 08:08:01.012838 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jl8c5" event={"ID":"8f99a056-2251-45ca-a55e-f970ab0fe1e2","Type":"ContainerStarted","Data":"f3e4727f4ba08177b8461aafc80be8621d5bf160338f88797cbb79336e866810"} Nov 24 08:08:01 crc kubenswrapper[4799]: I1124 08:08:01.036490 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-jl8c5" podStartSLOduration=4.224298396 podStartE2EDuration="19.036460594s" podCreationTimestamp="2025-11-24 08:07:42 +0000 UTC" firstStartedPulling="2025-11-24 08:07:44.128618078 +0000 UTC m=+4809.784600562" lastFinishedPulling="2025-11-24 08:07:58.940780266 +0000 UTC m=+4824.596762760" observedRunningTime="2025-11-24 08:08:01.029981484 +0000 UTC m=+4826.685963958" watchObservedRunningTime="2025-11-24 08:08:01.036460594 +0000 UTC m=+4826.692443088" Nov 24 08:08:04 crc kubenswrapper[4799]: I1124 08:08:04.048085 4799 generic.go:334] "Generic (PLEG): container finished" podID="8f99a056-2251-45ca-a55e-f970ab0fe1e2" containerID="f3e4727f4ba08177b8461aafc80be8621d5bf160338f88797cbb79336e866810" exitCode=0 Nov 24 08:08:04 crc kubenswrapper[4799]: I1124 08:08:04.048301 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jl8c5" event={"ID":"8f99a056-2251-45ca-a55e-f970ab0fe1e2","Type":"ContainerDied","Data":"f3e4727f4ba08177b8461aafc80be8621d5bf160338f88797cbb79336e866810"} Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.489699 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jl8c5" Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.502089 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-config-data\") pod \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.502288 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-combined-ca-bundle\") pod \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.502332 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-db-sync-config-data\") pod \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.502366 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgngg\" (UniqueName: \"kubernetes.io/projected/8f99a056-2251-45ca-a55e-f970ab0fe1e2-kube-api-access-zgngg\") pod \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\" (UID: \"8f99a056-2251-45ca-a55e-f970ab0fe1e2\") " Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.552027 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8f99a056-2251-45ca-a55e-f970ab0fe1e2" (UID: "8f99a056-2251-45ca-a55e-f970ab0fe1e2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.552135 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f99a056-2251-45ca-a55e-f970ab0fe1e2-kube-api-access-zgngg" (OuterVolumeSpecName: "kube-api-access-zgngg") pod "8f99a056-2251-45ca-a55e-f970ab0fe1e2" (UID: "8f99a056-2251-45ca-a55e-f970ab0fe1e2"). InnerVolumeSpecName "kube-api-access-zgngg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.555751 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f99a056-2251-45ca-a55e-f970ab0fe1e2" (UID: "8f99a056-2251-45ca-a55e-f970ab0fe1e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.587375 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-config-data" (OuterVolumeSpecName: "config-data") pod "8f99a056-2251-45ca-a55e-f970ab0fe1e2" (UID: "8f99a056-2251-45ca-a55e-f970ab0fe1e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.604739 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.604949 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.605057 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgngg\" (UniqueName: \"kubernetes.io/projected/8f99a056-2251-45ca-a55e-f970ab0fe1e2-kube-api-access-zgngg\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:05 crc kubenswrapper[4799]: I1124 08:08:05.605170 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f99a056-2251-45ca-a55e-f970ab0fe1e2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.072160 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jl8c5" event={"ID":"8f99a056-2251-45ca-a55e-f970ab0fe1e2","Type":"ContainerDied","Data":"741757f8935f0314af3ba153f1d0fbfbd255244bd5b4354963e49b52eba1f0fc"} Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.072220 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="741757f8935f0314af3ba153f1d0fbfbd255244bd5b4354963e49b52eba1f0fc" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.072233 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jl8c5" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.473749 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6664485f69-r9nj4"] Nov 24 08:08:06 crc kubenswrapper[4799]: E1124 08:08:06.474113 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22864cc4-2d09-49e6-b55c-d10295e62c8e" containerName="extract-content" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.474128 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="22864cc4-2d09-49e6-b55c-d10295e62c8e" containerName="extract-content" Nov 24 08:08:06 crc kubenswrapper[4799]: E1124 08:08:06.474141 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22864cc4-2d09-49e6-b55c-d10295e62c8e" containerName="registry-server" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.474147 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="22864cc4-2d09-49e6-b55c-d10295e62c8e" containerName="registry-server" Nov 24 08:08:06 crc kubenswrapper[4799]: E1124 08:08:06.474165 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22864cc4-2d09-49e6-b55c-d10295e62c8e" containerName="extract-utilities" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.474172 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="22864cc4-2d09-49e6-b55c-d10295e62c8e" containerName="extract-utilities" Nov 24 08:08:06 crc kubenswrapper[4799]: E1124 08:08:06.474189 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f99a056-2251-45ca-a55e-f970ab0fe1e2" containerName="glance-db-sync" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.474194 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f99a056-2251-45ca-a55e-f970ab0fe1e2" containerName="glance-db-sync" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.474358 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f99a056-2251-45ca-a55e-f970ab0fe1e2" containerName="glance-db-sync" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.474387 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="22864cc4-2d09-49e6-b55c-d10295e62c8e" containerName="registry-server" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.475230 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.491308 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.492894 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.499917 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.500098 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.500213 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.500435 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lwt7f" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.502449 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6664485f69-r9nj4"] Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.526554 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.621968 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.622055 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdkmk\" (UniqueName: \"kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-kube-api-access-xdkmk\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.622106 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-config\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.622124 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-logs\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.622214 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-sb\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.622275 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.622325 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-dns-svc\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.622371 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.622399 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.622456 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-ceph\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.622507 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-nb\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.622536 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pfd4\" (UniqueName: \"kubernetes.io/projected/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-kube-api-access-9pfd4\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.632981 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.634817 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.636639 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.651704 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.723729 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c69d\" (UniqueName: \"kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-kube-api-access-8c69d\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.723775 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.723800 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-ceph\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.723835 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.724596 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-nb\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.724632 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pfd4\" (UniqueName: \"kubernetes.io/projected/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-kube-api-access-9pfd4\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.724689 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.724747 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-ceph\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.724774 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdkmk\" (UniqueName: \"kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-kube-api-access-xdkmk\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.725583 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-nb\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.725597 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-config\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.725650 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-logs\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.726069 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-logs\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.726110 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-config\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.726160 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.726693 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-sb\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.726746 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-logs\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.726783 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.726823 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-dns-svc\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.726926 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.726971 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.727279 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.728305 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-dns-svc\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.728355 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.729240 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-ceph\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.730526 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.731556 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-sb\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.745766 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.746748 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.748162 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pfd4\" (UniqueName: \"kubernetes.io/projected/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-kube-api-access-9pfd4\") pod \"dnsmasq-dns-6664485f69-r9nj4\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.751443 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdkmk\" (UniqueName: \"kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-kube-api-access-xdkmk\") pod \"glance-default-external-api-0\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.794526 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.816303 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.829723 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.829786 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-logs\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.829840 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.829892 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c69d\" (UniqueName: \"kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-kube-api-access-8c69d\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.829918 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.830001 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.830071 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-ceph\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.830205 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-logs\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.830716 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.834013 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-ceph\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.837116 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.837450 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.838403 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.848653 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c69d\" (UniqueName: \"kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-kube-api-access-8c69d\") pod \"glance-default-internal-api-0\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:06 crc kubenswrapper[4799]: I1124 08:08:06.951675 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:07 crc kubenswrapper[4799]: I1124 08:08:07.245968 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6664485f69-r9nj4"] Nov 24 08:08:07 crc kubenswrapper[4799]: W1124 08:08:07.247879 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e8fc5fa_8dbe_4ce3_9c96_53e58164a532.slice/crio-44168d28414276d8ef58c7b1210b16c06e585fb50cfb430d96b4a2d0670f1da9 WatchSource:0}: Error finding container 44168d28414276d8ef58c7b1210b16c06e585fb50cfb430d96b4a2d0670f1da9: Status 404 returned error can't find the container with id 44168d28414276d8ef58c7b1210b16c06e585fb50cfb430d96b4a2d0670f1da9 Nov 24 08:08:07 crc kubenswrapper[4799]: I1124 08:08:07.407839 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:08:07 crc kubenswrapper[4799]: I1124 08:08:07.488401 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:08:07 crc kubenswrapper[4799]: I1124 08:08:07.598979 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:08:08 crc kubenswrapper[4799]: I1124 08:08:08.106265 4799 generic.go:334] "Generic (PLEG): container finished" podID="3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" containerID="bc3c6afbfe50f85305b930f45daae4a2e69cc4ab992b620b03793e52bffefb56" exitCode=0 Nov 24 08:08:08 crc kubenswrapper[4799]: I1124 08:08:08.106615 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" event={"ID":"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532","Type":"ContainerDied","Data":"bc3c6afbfe50f85305b930f45daae4a2e69cc4ab992b620b03793e52bffefb56"} Nov 24 08:08:08 crc kubenswrapper[4799]: I1124 08:08:08.106660 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" event={"ID":"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532","Type":"ContainerStarted","Data":"44168d28414276d8ef58c7b1210b16c06e585fb50cfb430d96b4a2d0670f1da9"} Nov 24 08:08:08 crc kubenswrapper[4799]: I1124 08:08:08.109562 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4329dec8-786f-44b2-8892-be0f6db031be","Type":"ContainerStarted","Data":"dce18a6efb55dea4454d2850aada3bfb138ceb650c15da2ed6d55d841680ad52"} Nov 24 08:08:08 crc kubenswrapper[4799]: I1124 08:08:08.111703 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5ec7f76-decd-486d-a91f-379fd96e691e","Type":"ContainerStarted","Data":"f63509a7412fca09354b5130fa5c6c72a4ba61a175411024c02a2846ea22dc23"} Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.132252 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" event={"ID":"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532","Type":"ContainerStarted","Data":"cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4"} Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.132671 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.135956 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4329dec8-786f-44b2-8892-be0f6db031be","Type":"ContainerStarted","Data":"c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304"} Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.136008 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4329dec8-786f-44b2-8892-be0f6db031be","Type":"ContainerStarted","Data":"92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6"} Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.138061 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5ec7f76-decd-486d-a91f-379fd96e691e","Type":"ContainerStarted","Data":"2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54"} Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.138168 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5ec7f76-decd-486d-a91f-379fd96e691e","Type":"ContainerStarted","Data":"c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe"} Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.138214 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c5ec7f76-decd-486d-a91f-379fd96e691e" containerName="glance-httpd" containerID="cri-o://c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe" gracePeriod=30 Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.138191 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c5ec7f76-decd-486d-a91f-379fd96e691e" containerName="glance-log" containerID="cri-o://2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54" gracePeriod=30 Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.159738 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" podStartSLOduration=3.159716344 podStartE2EDuration="3.159716344s" podCreationTimestamp="2025-11-24 08:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:08:09.152480223 +0000 UTC m=+4834.808462697" watchObservedRunningTime="2025-11-24 08:08:09.159716344 +0000 UTC m=+4834.815698818" Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.178533 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.178517786 podStartE2EDuration="3.178517786s" podCreationTimestamp="2025-11-24 08:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:08:09.173632671 +0000 UTC m=+4834.829615145" watchObservedRunningTime="2025-11-24 08:08:09.178517786 +0000 UTC m=+4834.834500260" Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.213170 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.2131533389999998 podStartE2EDuration="3.213153339s" podCreationTimestamp="2025-11-24 08:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:08:09.204490648 +0000 UTC m=+4834.860473122" watchObservedRunningTime="2025-11-24 08:08:09.213153339 +0000 UTC m=+4834.869135813" Nov 24 08:08:09 crc kubenswrapper[4799]: I1124 08:08:09.412093 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.088998 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.148731 4799 generic.go:334] "Generic (PLEG): container finished" podID="c5ec7f76-decd-486d-a91f-379fd96e691e" containerID="c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe" exitCode=0 Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.148774 4799 generic.go:334] "Generic (PLEG): container finished" podID="c5ec7f76-decd-486d-a91f-379fd96e691e" containerID="2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54" exitCode=143 Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.148820 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5ec7f76-decd-486d-a91f-379fd96e691e","Type":"ContainerDied","Data":"c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe"} Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.148889 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5ec7f76-decd-486d-a91f-379fd96e691e","Type":"ContainerDied","Data":"2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54"} Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.148904 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5ec7f76-decd-486d-a91f-379fd96e691e","Type":"ContainerDied","Data":"f63509a7412fca09354b5130fa5c6c72a4ba61a175411024c02a2846ea22dc23"} Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.148922 4799 scope.go:117] "RemoveContainer" containerID="c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.148922 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.171181 4799 scope.go:117] "RemoveContainer" containerID="2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.190906 4799 scope.go:117] "RemoveContainer" containerID="c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe" Nov 24 08:08:10 crc kubenswrapper[4799]: E1124 08:08:10.191388 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe\": container with ID starting with c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe not found: ID does not exist" containerID="c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.191429 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe"} err="failed to get container status \"c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe\": rpc error: code = NotFound desc = could not find container \"c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe\": container with ID starting with c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe not found: ID does not exist" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.191449 4799 scope.go:117] "RemoveContainer" containerID="2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54" Nov 24 08:08:10 crc kubenswrapper[4799]: E1124 08:08:10.191798 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54\": container with ID starting with 2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54 not found: ID does not exist" containerID="2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.191870 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54"} err="failed to get container status \"2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54\": rpc error: code = NotFound desc = could not find container \"2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54\": container with ID starting with 2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54 not found: ID does not exist" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.191904 4799 scope.go:117] "RemoveContainer" containerID="c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.192300 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe"} err="failed to get container status \"c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe\": rpc error: code = NotFound desc = could not find container \"c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe\": container with ID starting with c67b1eba7559a0dc0ec189176e9970ffc1d299c2c6d57eabb0f798df40b76ebe not found: ID does not exist" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.192328 4799 scope.go:117] "RemoveContainer" containerID="2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.192587 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54"} err="failed to get container status \"2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54\": rpc error: code = NotFound desc = could not find container \"2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54\": container with ID starting with 2f83db8a06dafeb92f4fe776795cd04ae9658672b59235acb4faab84df725d54 not found: ID does not exist" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.227801 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-ceph\") pod \"c5ec7f76-decd-486d-a91f-379fd96e691e\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.227887 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdkmk\" (UniqueName: \"kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-kube-api-access-xdkmk\") pod \"c5ec7f76-decd-486d-a91f-379fd96e691e\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.227907 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-combined-ca-bundle\") pod \"c5ec7f76-decd-486d-a91f-379fd96e691e\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.228198 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-config-data\") pod \"c5ec7f76-decd-486d-a91f-379fd96e691e\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.229049 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-httpd-run\") pod \"c5ec7f76-decd-486d-a91f-379fd96e691e\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.229140 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-scripts\") pod \"c5ec7f76-decd-486d-a91f-379fd96e691e\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.229220 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-logs\") pod \"c5ec7f76-decd-486d-a91f-379fd96e691e\" (UID: \"c5ec7f76-decd-486d-a91f-379fd96e691e\") " Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.229388 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c5ec7f76-decd-486d-a91f-379fd96e691e" (UID: "c5ec7f76-decd-486d-a91f-379fd96e691e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.229608 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.229809 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-logs" (OuterVolumeSpecName: "logs") pod "c5ec7f76-decd-486d-a91f-379fd96e691e" (UID: "c5ec7f76-decd-486d-a91f-379fd96e691e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.237751 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-scripts" (OuterVolumeSpecName: "scripts") pod "c5ec7f76-decd-486d-a91f-379fd96e691e" (UID: "c5ec7f76-decd-486d-a91f-379fd96e691e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.237824 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-ceph" (OuterVolumeSpecName: "ceph") pod "c5ec7f76-decd-486d-a91f-379fd96e691e" (UID: "c5ec7f76-decd-486d-a91f-379fd96e691e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.238505 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-kube-api-access-xdkmk" (OuterVolumeSpecName: "kube-api-access-xdkmk") pod "c5ec7f76-decd-486d-a91f-379fd96e691e" (UID: "c5ec7f76-decd-486d-a91f-379fd96e691e"). InnerVolumeSpecName "kube-api-access-xdkmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.254267 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5ec7f76-decd-486d-a91f-379fd96e691e" (UID: "c5ec7f76-decd-486d-a91f-379fd96e691e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.273590 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-config-data" (OuterVolumeSpecName: "config-data") pod "c5ec7f76-decd-486d-a91f-379fd96e691e" (UID: "c5ec7f76-decd-486d-a91f-379fd96e691e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.333003 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdkmk\" (UniqueName: \"kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-kube-api-access-xdkmk\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.333040 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.333050 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.333062 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5ec7f76-decd-486d-a91f-379fd96e691e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.333071 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5ec7f76-decd-486d-a91f-379fd96e691e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.333078 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c5ec7f76-decd-486d-a91f-379fd96e691e-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.490173 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.513639 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.527083 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:08:10 crc kubenswrapper[4799]: E1124 08:08:10.527644 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ec7f76-decd-486d-a91f-379fd96e691e" containerName="glance-log" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.527673 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ec7f76-decd-486d-a91f-379fd96e691e" containerName="glance-log" Nov 24 08:08:10 crc kubenswrapper[4799]: E1124 08:08:10.527705 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ec7f76-decd-486d-a91f-379fd96e691e" containerName="glance-httpd" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.527717 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ec7f76-decd-486d-a91f-379fd96e691e" containerName="glance-httpd" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.528039 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ec7f76-decd-486d-a91f-379fd96e691e" containerName="glance-log" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.528091 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ec7f76-decd-486d-a91f-379fd96e691e" containerName="glance-httpd" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.529581 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.532391 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.532755 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.536746 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.536791 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-logs\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.536823 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-config-data\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.536899 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvx8q\" (UniqueName: \"kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-kube-api-access-jvx8q\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.536958 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.536985 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-ceph\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.537032 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-scripts\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.637781 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-scripts\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.637876 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.637918 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-logs\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.638086 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-config-data\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.638251 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvx8q\" (UniqueName: \"kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-kube-api-access-jvx8q\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.638286 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-logs\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.639239 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.639366 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-ceph\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.639469 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.642151 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-scripts\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.642674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-config-data\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.644888 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.645307 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-ceph\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.661799 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvx8q\" (UniqueName: \"kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-kube-api-access-jvx8q\") pod \"glance-default-external-api-0\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " pod="openstack/glance-default-external-api-0" Nov 24 08:08:10 crc kubenswrapper[4799]: I1124 08:08:10.885920 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 08:08:11 crc kubenswrapper[4799]: I1124 08:08:11.160864 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4329dec8-786f-44b2-8892-be0f6db031be" containerName="glance-log" containerID="cri-o://92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6" gracePeriod=30 Nov 24 08:08:11 crc kubenswrapper[4799]: I1124 08:08:11.161064 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4329dec8-786f-44b2-8892-be0f6db031be" containerName="glance-httpd" containerID="cri-o://c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304" gracePeriod=30 Nov 24 08:08:11 crc kubenswrapper[4799]: I1124 08:08:11.439464 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:08:11 crc kubenswrapper[4799]: W1124 08:08:11.452310 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod363dee68_bcb4_45f0_a16c_ab6a4d0d2b42.slice/crio-3c2f8f15e9b9b6f8f29b3cccdd0124a600ab7aa8030cbb00c90952409b212812 WatchSource:0}: Error finding container 3c2f8f15e9b9b6f8f29b3cccdd0124a600ab7aa8030cbb00c90952409b212812: Status 404 returned error can't find the container with id 3c2f8f15e9b9b6f8f29b3cccdd0124a600ab7aa8030cbb00c90952409b212812 Nov 24 08:08:11 crc kubenswrapper[4799]: I1124 08:08:11.640392 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5ec7f76-decd-486d-a91f-379fd96e691e" path="/var/lib/kubelet/pods/c5ec7f76-decd-486d-a91f-379fd96e691e/volumes" Nov 24 08:08:11 crc kubenswrapper[4799]: I1124 08:08:11.932448 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.065056 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-httpd-run\") pod \"4329dec8-786f-44b2-8892-be0f6db031be\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.065101 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-logs\") pod \"4329dec8-786f-44b2-8892-be0f6db031be\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.065150 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-combined-ca-bundle\") pod \"4329dec8-786f-44b2-8892-be0f6db031be\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.065395 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4329dec8-786f-44b2-8892-be0f6db031be" (UID: "4329dec8-786f-44b2-8892-be0f6db031be"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.065670 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-logs" (OuterVolumeSpecName: "logs") pod "4329dec8-786f-44b2-8892-be0f6db031be" (UID: "4329dec8-786f-44b2-8892-be0f6db031be"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.065169 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c69d\" (UniqueName: \"kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-kube-api-access-8c69d\") pod \"4329dec8-786f-44b2-8892-be0f6db031be\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.068891 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-ceph\") pod \"4329dec8-786f-44b2-8892-be0f6db031be\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.068954 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-scripts\") pod \"4329dec8-786f-44b2-8892-be0f6db031be\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.069031 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-config-data\") pod \"4329dec8-786f-44b2-8892-be0f6db031be\" (UID: \"4329dec8-786f-44b2-8892-be0f6db031be\") " Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.070010 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.070032 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4329dec8-786f-44b2-8892-be0f6db031be-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.073228 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-kube-api-access-8c69d" (OuterVolumeSpecName: "kube-api-access-8c69d") pod "4329dec8-786f-44b2-8892-be0f6db031be" (UID: "4329dec8-786f-44b2-8892-be0f6db031be"). InnerVolumeSpecName "kube-api-access-8c69d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.073335 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-scripts" (OuterVolumeSpecName: "scripts") pod "4329dec8-786f-44b2-8892-be0f6db031be" (UID: "4329dec8-786f-44b2-8892-be0f6db031be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.078069 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-ceph" (OuterVolumeSpecName: "ceph") pod "4329dec8-786f-44b2-8892-be0f6db031be" (UID: "4329dec8-786f-44b2-8892-be0f6db031be"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.101008 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4329dec8-786f-44b2-8892-be0f6db031be" (UID: "4329dec8-786f-44b2-8892-be0f6db031be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.154167 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-config-data" (OuterVolumeSpecName: "config-data") pod "4329dec8-786f-44b2-8892-be0f6db031be" (UID: "4329dec8-786f-44b2-8892-be0f6db031be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.172465 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42","Type":"ContainerStarted","Data":"3c2f8f15e9b9b6f8f29b3cccdd0124a600ab7aa8030cbb00c90952409b212812"} Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.174709 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.175377 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.175409 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.175435 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4329dec8-786f-44b2-8892-be0f6db031be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.175448 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c69d\" (UniqueName: \"kubernetes.io/projected/4329dec8-786f-44b2-8892-be0f6db031be-kube-api-access-8c69d\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.178199 4799 generic.go:334] "Generic (PLEG): container finished" podID="4329dec8-786f-44b2-8892-be0f6db031be" containerID="c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304" exitCode=0 Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.178224 4799 generic.go:334] "Generic (PLEG): container finished" podID="4329dec8-786f-44b2-8892-be0f6db031be" containerID="92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6" exitCode=143 Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.178238 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4329dec8-786f-44b2-8892-be0f6db031be","Type":"ContainerDied","Data":"c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304"} Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.178278 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4329dec8-786f-44b2-8892-be0f6db031be","Type":"ContainerDied","Data":"92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6"} Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.178290 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4329dec8-786f-44b2-8892-be0f6db031be","Type":"ContainerDied","Data":"dce18a6efb55dea4454d2850aada3bfb138ceb650c15da2ed6d55d841680ad52"} Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.178306 4799 scope.go:117] "RemoveContainer" containerID="c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.178444 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.245942 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.251268 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.268501 4799 scope.go:117] "RemoveContainer" containerID="92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.269739 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:08:12 crc kubenswrapper[4799]: E1124 08:08:12.270220 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4329dec8-786f-44b2-8892-be0f6db031be" containerName="glance-httpd" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.270238 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4329dec8-786f-44b2-8892-be0f6db031be" containerName="glance-httpd" Nov 24 08:08:12 crc kubenswrapper[4799]: E1124 08:08:12.270261 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4329dec8-786f-44b2-8892-be0f6db031be" containerName="glance-log" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.270271 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4329dec8-786f-44b2-8892-be0f6db031be" containerName="glance-log" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.270514 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4329dec8-786f-44b2-8892-be0f6db031be" containerName="glance-log" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.270533 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4329dec8-786f-44b2-8892-be0f6db031be" containerName="glance-httpd" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.271675 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.277268 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.287381 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.298819 4799 scope.go:117] "RemoveContainer" containerID="c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304" Nov 24 08:08:12 crc kubenswrapper[4799]: E1124 08:08:12.311705 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304\": container with ID starting with c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304 not found: ID does not exist" containerID="c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.311768 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304"} err="failed to get container status \"c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304\": rpc error: code = NotFound desc = could not find container \"c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304\": container with ID starting with c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304 not found: ID does not exist" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.311801 4799 scope.go:117] "RemoveContainer" containerID="92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6" Nov 24 08:08:12 crc kubenswrapper[4799]: E1124 08:08:12.312620 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6\": container with ID starting with 92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6 not found: ID does not exist" containerID="92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.312665 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6"} err="failed to get container status \"92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6\": rpc error: code = NotFound desc = could not find container \"92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6\": container with ID starting with 92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6 not found: ID does not exist" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.312691 4799 scope.go:117] "RemoveContainer" containerID="c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.313115 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304"} err="failed to get container status \"c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304\": rpc error: code = NotFound desc = could not find container \"c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304\": container with ID starting with c063c6493ccdde034e765817c85134135c4a211a8173549aec0fa366b4cfe304 not found: ID does not exist" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.313153 4799 scope.go:117] "RemoveContainer" containerID="92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.315819 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6"} err="failed to get container status \"92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6\": rpc error: code = NotFound desc = could not find container \"92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6\": container with ID starting with 92ab81917b76c56d71acc19ab93120e58000a346ceda20e752720185de8c10a6 not found: ID does not exist" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.378577 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-ceph\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.378629 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slkkn\" (UniqueName: \"kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-kube-api-access-slkkn\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.378678 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.378717 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.378770 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.378798 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-logs\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.378815 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.481158 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.481223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-logs\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.481242 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.481277 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-ceph\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.481295 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slkkn\" (UniqueName: \"kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-kube-api-access-slkkn\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.481331 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.481363 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.482152 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.483290 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-logs\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.486268 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.486626 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.487444 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-ceph\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.488584 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.503521 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slkkn\" (UniqueName: \"kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-kube-api-access-slkkn\") pod \"glance-default-internal-api-0\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:08:12 crc kubenswrapper[4799]: I1124 08:08:12.594176 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:13 crc kubenswrapper[4799]: I1124 08:08:13.160128 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:08:13 crc kubenswrapper[4799]: W1124 08:08:13.170652 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46b45ccf_b3d0_44da_a16e_8d1c5a8c3949.slice/crio-0e71411b4c7933effbdf63bfe8b4f4b18ecbab01ac7bf9ce362c276d9f96dd2e WatchSource:0}: Error finding container 0e71411b4c7933effbdf63bfe8b4f4b18ecbab01ac7bf9ce362c276d9f96dd2e: Status 404 returned error can't find the container with id 0e71411b4c7933effbdf63bfe8b4f4b18ecbab01ac7bf9ce362c276d9f96dd2e Nov 24 08:08:13 crc kubenswrapper[4799]: I1124 08:08:13.189900 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42","Type":"ContainerStarted","Data":"3b8326afd0a897e5ef200985d1454222061c289ca54efbb6114e721972fbdeed"} Nov 24 08:08:13 crc kubenswrapper[4799]: I1124 08:08:13.189952 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42","Type":"ContainerStarted","Data":"dc14a6488fdb8f0b78a111c5df96c45942e6965d8020be92ae627a6559759dac"} Nov 24 08:08:13 crc kubenswrapper[4799]: I1124 08:08:13.191305 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949","Type":"ContainerStarted","Data":"0e71411b4c7933effbdf63bfe8b4f4b18ecbab01ac7bf9ce362c276d9f96dd2e"} Nov 24 08:08:13 crc kubenswrapper[4799]: I1124 08:08:13.217911 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.217894479 podStartE2EDuration="3.217894479s" podCreationTimestamp="2025-11-24 08:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:08:13.211295295 +0000 UTC m=+4838.867277789" watchObservedRunningTime="2025-11-24 08:08:13.217894479 +0000 UTC m=+4838.873876953" Nov 24 08:08:13 crc kubenswrapper[4799]: I1124 08:08:13.638633 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4329dec8-786f-44b2-8892-be0f6db031be" path="/var/lib/kubelet/pods/4329dec8-786f-44b2-8892-be0f6db031be/volumes" Nov 24 08:08:14 crc kubenswrapper[4799]: I1124 08:08:14.209511 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949","Type":"ContainerStarted","Data":"c1939c2300d3acc7a7b5dcf826e0a316e19046ae2e5527f9484594291f376394"} Nov 24 08:08:15 crc kubenswrapper[4799]: I1124 08:08:15.239836 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949","Type":"ContainerStarted","Data":"9baa784af9346bce45134500375430672747e10413316507b7bed0357ee4d7d8"} Nov 24 08:08:15 crc kubenswrapper[4799]: I1124 08:08:15.278075 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.27805184 podStartE2EDuration="3.27805184s" podCreationTimestamp="2025-11-24 08:08:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:08:15.273208305 +0000 UTC m=+4840.929190789" watchObservedRunningTime="2025-11-24 08:08:15.27805184 +0000 UTC m=+4840.934034324" Nov 24 08:08:16 crc kubenswrapper[4799]: I1124 08:08:16.796049 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:16 crc kubenswrapper[4799]: I1124 08:08:16.912264 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-767b5748b7-mwgqh"] Nov 24 08:08:16 crc kubenswrapper[4799]: I1124 08:08:16.912535 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" podUID="71659c8f-1be8-4030-b8c4-c1ad3b3e850a" containerName="dnsmasq-dns" containerID="cri-o://09631b72971a6b9d5fed65d4c752c45991cb93f1056d637fd3b470e43ece9247" gracePeriod=10 Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.262238 4799 generic.go:334] "Generic (PLEG): container finished" podID="71659c8f-1be8-4030-b8c4-c1ad3b3e850a" containerID="09631b72971a6b9d5fed65d4c752c45991cb93f1056d637fd3b470e43ece9247" exitCode=0 Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.262333 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" event={"ID":"71659c8f-1be8-4030-b8c4-c1ad3b3e850a","Type":"ContainerDied","Data":"09631b72971a6b9d5fed65d4c752c45991cb93f1056d637fd3b470e43ece9247"} Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.386683 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.581310 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-config\") pod \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.581441 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-sb\") pod \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.581499 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rpvh\" (UniqueName: \"kubernetes.io/projected/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-kube-api-access-8rpvh\") pod \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.581538 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-dns-svc\") pod \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.581557 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-nb\") pod \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\" (UID: \"71659c8f-1be8-4030-b8c4-c1ad3b3e850a\") " Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.591198 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-kube-api-access-8rpvh" (OuterVolumeSpecName: "kube-api-access-8rpvh") pod "71659c8f-1be8-4030-b8c4-c1ad3b3e850a" (UID: "71659c8f-1be8-4030-b8c4-c1ad3b3e850a"). InnerVolumeSpecName "kube-api-access-8rpvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.622765 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "71659c8f-1be8-4030-b8c4-c1ad3b3e850a" (UID: "71659c8f-1be8-4030-b8c4-c1ad3b3e850a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.627442 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "71659c8f-1be8-4030-b8c4-c1ad3b3e850a" (UID: "71659c8f-1be8-4030-b8c4-c1ad3b3e850a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.635442 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-config" (OuterVolumeSpecName: "config") pod "71659c8f-1be8-4030-b8c4-c1ad3b3e850a" (UID: "71659c8f-1be8-4030-b8c4-c1ad3b3e850a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.649986 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "71659c8f-1be8-4030-b8c4-c1ad3b3e850a" (UID: "71659c8f-1be8-4030-b8c4-c1ad3b3e850a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.683433 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.683476 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.683486 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rpvh\" (UniqueName: \"kubernetes.io/projected/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-kube-api-access-8rpvh\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.683496 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:17 crc kubenswrapper[4799]: I1124 08:08:17.683505 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71659c8f-1be8-4030-b8c4-c1ad3b3e850a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:18 crc kubenswrapper[4799]: I1124 08:08:18.278065 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" event={"ID":"71659c8f-1be8-4030-b8c4-c1ad3b3e850a","Type":"ContainerDied","Data":"02c89c179f71174c20e5e5006059ebcc377f6b404c164a26a25a139e269cc9f2"} Nov 24 08:08:18 crc kubenswrapper[4799]: I1124 08:08:18.278586 4799 scope.go:117] "RemoveContainer" containerID="09631b72971a6b9d5fed65d4c752c45991cb93f1056d637fd3b470e43ece9247" Nov 24 08:08:18 crc kubenswrapper[4799]: I1124 08:08:18.278129 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-767b5748b7-mwgqh" Nov 24 08:08:18 crc kubenswrapper[4799]: I1124 08:08:18.300220 4799 scope.go:117] "RemoveContainer" containerID="c90bf2f1adf577afb3c7c2df06a2a01c096bb24e43a2922959e3c5de774bdf39" Nov 24 08:08:18 crc kubenswrapper[4799]: I1124 08:08:18.318373 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-767b5748b7-mwgqh"] Nov 24 08:08:18 crc kubenswrapper[4799]: I1124 08:08:18.325968 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-767b5748b7-mwgqh"] Nov 24 08:08:19 crc kubenswrapper[4799]: I1124 08:08:19.641531 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71659c8f-1be8-4030-b8c4-c1ad3b3e850a" path="/var/lib/kubelet/pods/71659c8f-1be8-4030-b8c4-c1ad3b3e850a/volumes" Nov 24 08:08:20 crc kubenswrapper[4799]: I1124 08:08:20.886656 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 08:08:20 crc kubenswrapper[4799]: I1124 08:08:20.887159 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 08:08:20 crc kubenswrapper[4799]: I1124 08:08:20.917387 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 08:08:20 crc kubenswrapper[4799]: I1124 08:08:20.935062 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 08:08:21 crc kubenswrapper[4799]: I1124 08:08:21.328248 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 08:08:21 crc kubenswrapper[4799]: I1124 08:08:21.328284 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 08:08:22 crc kubenswrapper[4799]: I1124 08:08:22.595237 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:22 crc kubenswrapper[4799]: I1124 08:08:22.596343 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:22 crc kubenswrapper[4799]: I1124 08:08:22.628984 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:22 crc kubenswrapper[4799]: I1124 08:08:22.637704 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:23 crc kubenswrapper[4799]: I1124 08:08:23.345066 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:23 crc kubenswrapper[4799]: I1124 08:08:23.345124 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:23 crc kubenswrapper[4799]: I1124 08:08:23.376974 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 08:08:23 crc kubenswrapper[4799]: I1124 08:08:23.377052 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:08:23 crc kubenswrapper[4799]: I1124 08:08:23.377486 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 08:08:25 crc kubenswrapper[4799]: I1124 08:08:25.330460 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:25 crc kubenswrapper[4799]: I1124 08:08:25.340342 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.356352 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-pfqpq"] Nov 24 08:08:33 crc kubenswrapper[4799]: E1124 08:08:33.357679 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71659c8f-1be8-4030-b8c4-c1ad3b3e850a" containerName="init" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.357701 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="71659c8f-1be8-4030-b8c4-c1ad3b3e850a" containerName="init" Nov 24 08:08:33 crc kubenswrapper[4799]: E1124 08:08:33.357733 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71659c8f-1be8-4030-b8c4-c1ad3b3e850a" containerName="dnsmasq-dns" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.357742 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="71659c8f-1be8-4030-b8c4-c1ad3b3e850a" containerName="dnsmasq-dns" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.358130 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="71659c8f-1be8-4030-b8c4-c1ad3b3e850a" containerName="dnsmasq-dns" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.359189 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pfqpq" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.375937 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-pfqpq"] Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.433925 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-0b82-account-create-bw62n"] Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.435256 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0b82-account-create-bw62n" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.437036 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.451262 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0b82-account-create-bw62n"] Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.464917 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdgsn\" (UniqueName: \"kubernetes.io/projected/511a70ee-d235-48a3-813b-956c39460f01-kube-api-access-rdgsn\") pod \"placement-db-create-pfqpq\" (UID: \"511a70ee-d235-48a3-813b-956c39460f01\") " pod="openstack/placement-db-create-pfqpq" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.464986 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/511a70ee-d235-48a3-813b-956c39460f01-operator-scripts\") pod \"placement-db-create-pfqpq\" (UID: \"511a70ee-d235-48a3-813b-956c39460f01\") " pod="openstack/placement-db-create-pfqpq" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.566728 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdgsn\" (UniqueName: \"kubernetes.io/projected/511a70ee-d235-48a3-813b-956c39460f01-kube-api-access-rdgsn\") pod \"placement-db-create-pfqpq\" (UID: \"511a70ee-d235-48a3-813b-956c39460f01\") " pod="openstack/placement-db-create-pfqpq" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.566836 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/511a70ee-d235-48a3-813b-956c39460f01-operator-scripts\") pod \"placement-db-create-pfqpq\" (UID: \"511a70ee-d235-48a3-813b-956c39460f01\") " pod="openstack/placement-db-create-pfqpq" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.566983 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-operator-scripts\") pod \"placement-0b82-account-create-bw62n\" (UID: \"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35\") " pod="openstack/placement-0b82-account-create-bw62n" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.567039 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prfmd\" (UniqueName: \"kubernetes.io/projected/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-kube-api-access-prfmd\") pod \"placement-0b82-account-create-bw62n\" (UID: \"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35\") " pod="openstack/placement-0b82-account-create-bw62n" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.567963 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/511a70ee-d235-48a3-813b-956c39460f01-operator-scripts\") pod \"placement-db-create-pfqpq\" (UID: \"511a70ee-d235-48a3-813b-956c39460f01\") " pod="openstack/placement-db-create-pfqpq" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.584999 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdgsn\" (UniqueName: \"kubernetes.io/projected/511a70ee-d235-48a3-813b-956c39460f01-kube-api-access-rdgsn\") pod \"placement-db-create-pfqpq\" (UID: \"511a70ee-d235-48a3-813b-956c39460f01\") " pod="openstack/placement-db-create-pfqpq" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.669183 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-operator-scripts\") pod \"placement-0b82-account-create-bw62n\" (UID: \"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35\") " pod="openstack/placement-0b82-account-create-bw62n" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.669483 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prfmd\" (UniqueName: \"kubernetes.io/projected/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-kube-api-access-prfmd\") pod \"placement-0b82-account-create-bw62n\" (UID: \"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35\") " pod="openstack/placement-0b82-account-create-bw62n" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.670082 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-operator-scripts\") pod \"placement-0b82-account-create-bw62n\" (UID: \"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35\") " pod="openstack/placement-0b82-account-create-bw62n" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.690282 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prfmd\" (UniqueName: \"kubernetes.io/projected/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-kube-api-access-prfmd\") pod \"placement-0b82-account-create-bw62n\" (UID: \"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35\") " pod="openstack/placement-0b82-account-create-bw62n" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.691787 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pfqpq" Nov 24 08:08:33 crc kubenswrapper[4799]: I1124 08:08:33.756498 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0b82-account-create-bw62n" Nov 24 08:08:34 crc kubenswrapper[4799]: I1124 08:08:34.121185 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-pfqpq"] Nov 24 08:08:34 crc kubenswrapper[4799]: I1124 08:08:34.270002 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0b82-account-create-bw62n"] Nov 24 08:08:34 crc kubenswrapper[4799]: W1124 08:08:34.277499 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8d4ba69_d6d5_434d_a20b_f87c4f3e8b35.slice/crio-c9af5505f118ce9904c950c2ad3e7b347c9a5f58f75ba027622b3698d3b64d20 WatchSource:0}: Error finding container c9af5505f118ce9904c950c2ad3e7b347c9a5f58f75ba027622b3698d3b64d20: Status 404 returned error can't find the container with id c9af5505f118ce9904c950c2ad3e7b347c9a5f58f75ba027622b3698d3b64d20 Nov 24 08:08:34 crc kubenswrapper[4799]: I1124 08:08:34.457470 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0b82-account-create-bw62n" event={"ID":"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35","Type":"ContainerStarted","Data":"22bb21315f2a8566eadbc6d049042a8f6fc00e6b9e79af02149377c148f6dd7d"} Nov 24 08:08:34 crc kubenswrapper[4799]: I1124 08:08:34.457516 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0b82-account-create-bw62n" event={"ID":"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35","Type":"ContainerStarted","Data":"c9af5505f118ce9904c950c2ad3e7b347c9a5f58f75ba027622b3698d3b64d20"} Nov 24 08:08:34 crc kubenswrapper[4799]: I1124 08:08:34.458978 4799 generic.go:334] "Generic (PLEG): container finished" podID="511a70ee-d235-48a3-813b-956c39460f01" containerID="05f6af4fec51a400e21d3d540f4c963b3f3946bbec60ea13ccf30c96220851c6" exitCode=0 Nov 24 08:08:34 crc kubenswrapper[4799]: I1124 08:08:34.459041 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pfqpq" event={"ID":"511a70ee-d235-48a3-813b-956c39460f01","Type":"ContainerDied","Data":"05f6af4fec51a400e21d3d540f4c963b3f3946bbec60ea13ccf30c96220851c6"} Nov 24 08:08:34 crc kubenswrapper[4799]: I1124 08:08:34.459066 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pfqpq" event={"ID":"511a70ee-d235-48a3-813b-956c39460f01","Type":"ContainerStarted","Data":"7c6bfc3e5c9483e1e376389da1de867789949ef03a880f8f7f6ed4b3c5e786de"} Nov 24 08:08:34 crc kubenswrapper[4799]: I1124 08:08:34.473566 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-0b82-account-create-bw62n" podStartSLOduration=1.473545726 podStartE2EDuration="1.473545726s" podCreationTimestamp="2025-11-24 08:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:08:34.470006748 +0000 UTC m=+4860.125989222" watchObservedRunningTime="2025-11-24 08:08:34.473545726 +0000 UTC m=+4860.129528200" Nov 24 08:08:35 crc kubenswrapper[4799]: I1124 08:08:35.469582 4799 generic.go:334] "Generic (PLEG): container finished" podID="e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35" containerID="22bb21315f2a8566eadbc6d049042a8f6fc00e6b9e79af02149377c148f6dd7d" exitCode=0 Nov 24 08:08:35 crc kubenswrapper[4799]: I1124 08:08:35.469679 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0b82-account-create-bw62n" event={"ID":"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35","Type":"ContainerDied","Data":"22bb21315f2a8566eadbc6d049042a8f6fc00e6b9e79af02149377c148f6dd7d"} Nov 24 08:08:35 crc kubenswrapper[4799]: I1124 08:08:35.854902 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pfqpq" Nov 24 08:08:36 crc kubenswrapper[4799]: I1124 08:08:36.009266 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/511a70ee-d235-48a3-813b-956c39460f01-operator-scripts\") pod \"511a70ee-d235-48a3-813b-956c39460f01\" (UID: \"511a70ee-d235-48a3-813b-956c39460f01\") " Nov 24 08:08:36 crc kubenswrapper[4799]: I1124 08:08:36.009582 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdgsn\" (UniqueName: \"kubernetes.io/projected/511a70ee-d235-48a3-813b-956c39460f01-kube-api-access-rdgsn\") pod \"511a70ee-d235-48a3-813b-956c39460f01\" (UID: \"511a70ee-d235-48a3-813b-956c39460f01\") " Nov 24 08:08:36 crc kubenswrapper[4799]: I1124 08:08:36.010101 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/511a70ee-d235-48a3-813b-956c39460f01-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "511a70ee-d235-48a3-813b-956c39460f01" (UID: "511a70ee-d235-48a3-813b-956c39460f01"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:08:36 crc kubenswrapper[4799]: I1124 08:08:36.010552 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/511a70ee-d235-48a3-813b-956c39460f01-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:36 crc kubenswrapper[4799]: I1124 08:08:36.015289 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/511a70ee-d235-48a3-813b-956c39460f01-kube-api-access-rdgsn" (OuterVolumeSpecName: "kube-api-access-rdgsn") pod "511a70ee-d235-48a3-813b-956c39460f01" (UID: "511a70ee-d235-48a3-813b-956c39460f01"). InnerVolumeSpecName "kube-api-access-rdgsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:08:36 crc kubenswrapper[4799]: I1124 08:08:36.112548 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdgsn\" (UniqueName: \"kubernetes.io/projected/511a70ee-d235-48a3-813b-956c39460f01-kube-api-access-rdgsn\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:36 crc kubenswrapper[4799]: I1124 08:08:36.499878 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pfqpq" Nov 24 08:08:36 crc kubenswrapper[4799]: I1124 08:08:36.499922 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pfqpq" event={"ID":"511a70ee-d235-48a3-813b-956c39460f01","Type":"ContainerDied","Data":"7c6bfc3e5c9483e1e376389da1de867789949ef03a880f8f7f6ed4b3c5e786de"} Nov 24 08:08:36 crc kubenswrapper[4799]: I1124 08:08:36.499975 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c6bfc3e5c9483e1e376389da1de867789949ef03a880f8f7f6ed4b3c5e786de" Nov 24 08:08:36 crc kubenswrapper[4799]: I1124 08:08:36.871978 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0b82-account-create-bw62n" Nov 24 08:08:37 crc kubenswrapper[4799]: I1124 08:08:37.028080 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-operator-scripts\") pod \"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35\" (UID: \"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35\") " Nov 24 08:08:37 crc kubenswrapper[4799]: I1124 08:08:37.028400 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prfmd\" (UniqueName: \"kubernetes.io/projected/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-kube-api-access-prfmd\") pod \"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35\" (UID: \"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35\") " Nov 24 08:08:37 crc kubenswrapper[4799]: I1124 08:08:37.028975 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35" (UID: "e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:08:37 crc kubenswrapper[4799]: I1124 08:08:37.030030 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:37 crc kubenswrapper[4799]: I1124 08:08:37.033152 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-kube-api-access-prfmd" (OuterVolumeSpecName: "kube-api-access-prfmd") pod "e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35" (UID: "e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35"). InnerVolumeSpecName "kube-api-access-prfmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:08:37 crc kubenswrapper[4799]: I1124 08:08:37.131803 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prfmd\" (UniqueName: \"kubernetes.io/projected/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35-kube-api-access-prfmd\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:37 crc kubenswrapper[4799]: I1124 08:08:37.512140 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0b82-account-create-bw62n" event={"ID":"e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35","Type":"ContainerDied","Data":"c9af5505f118ce9904c950c2ad3e7b347c9a5f58f75ba027622b3698d3b64d20"} Nov 24 08:08:37 crc kubenswrapper[4799]: I1124 08:08:37.512179 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9af5505f118ce9904c950c2ad3e7b347c9a5f58f75ba027622b3698d3b64d20" Nov 24 08:08:37 crc kubenswrapper[4799]: I1124 08:08:37.513072 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0b82-account-create-bw62n" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.788631 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c5dc9fb77-h8l9n"] Nov 24 08:08:38 crc kubenswrapper[4799]: E1124 08:08:38.790883 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35" containerName="mariadb-account-create" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.791006 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35" containerName="mariadb-account-create" Nov 24 08:08:38 crc kubenswrapper[4799]: E1124 08:08:38.791111 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="511a70ee-d235-48a3-813b-956c39460f01" containerName="mariadb-database-create" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.791203 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="511a70ee-d235-48a3-813b-956c39460f01" containerName="mariadb-database-create" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.791501 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="511a70ee-d235-48a3-813b-956c39460f01" containerName="mariadb-database-create" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.791591 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35" containerName="mariadb-account-create" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.792802 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.817240 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5dc9fb77-h8l9n"] Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.829103 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-78cxl"] Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.830629 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.832659 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-p8cg9" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.836384 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.836545 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.849583 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-78cxl"] Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.963389 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85pb4\" (UniqueName: \"kubernetes.io/projected/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-kube-api-access-85pb4\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.963467 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6698e591-c23a-477e-890e-cddb65f66fad-logs\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.963504 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-combined-ca-bundle\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.963567 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-config-data\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.963598 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-dns-svc\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.963646 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.963677 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-config\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.963833 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m2cv\" (UniqueName: \"kubernetes.io/projected/6698e591-c23a-477e-890e-cddb65f66fad-kube-api-access-2m2cv\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.963895 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:38 crc kubenswrapper[4799]: I1124 08:08:38.963959 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-scripts\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.065193 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.065260 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-config\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.065306 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m2cv\" (UniqueName: \"kubernetes.io/projected/6698e591-c23a-477e-890e-cddb65f66fad-kube-api-access-2m2cv\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.065329 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.065358 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-scripts\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.065399 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85pb4\" (UniqueName: \"kubernetes.io/projected/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-kube-api-access-85pb4\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.065424 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6698e591-c23a-477e-890e-cddb65f66fad-logs\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.065457 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-combined-ca-bundle\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.065530 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-config-data\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.065565 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-dns-svc\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.067121 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6698e591-c23a-477e-890e-cddb65f66fad-logs\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.067349 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.067418 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-dns-svc\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.067484 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.067607 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-config\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.071339 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-combined-ca-bundle\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.071744 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-scripts\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.081637 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-config-data\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.083389 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m2cv\" (UniqueName: \"kubernetes.io/projected/6698e591-c23a-477e-890e-cddb65f66fad-kube-api-access-2m2cv\") pod \"placement-db-sync-78cxl\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.085686 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85pb4\" (UniqueName: \"kubernetes.io/projected/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-kube-api-access-85pb4\") pod \"dnsmasq-dns-6c5dc9fb77-h8l9n\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.125227 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.158626 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:39 crc kubenswrapper[4799]: W1124 08:08:39.869463 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f8cae7e_b65e_4104_86b8_f2c6fa283f35.slice/crio-8ba29457ddfa81ff4b8f69ee286bd5612b88ba4b2b1a632bf3575978aa45483e WatchSource:0}: Error finding container 8ba29457ddfa81ff4b8f69ee286bd5612b88ba4b2b1a632bf3575978aa45483e: Status 404 returned error can't find the container with id 8ba29457ddfa81ff4b8f69ee286bd5612b88ba4b2b1a632bf3575978aa45483e Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.873882 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5dc9fb77-h8l9n"] Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.969523 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-78cxl"] Nov 24 08:08:39 crc kubenswrapper[4799]: W1124 08:08:39.970687 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6698e591_c23a_477e_890e_cddb65f66fad.slice/crio-b17c470d7411aafea8ddce8c46e6141e3a671cbb9c8054df93e487d78986a320 WatchSource:0}: Error finding container b17c470d7411aafea8ddce8c46e6141e3a671cbb9c8054df93e487d78986a320: Status 404 returned error can't find the container with id b17c470d7411aafea8ddce8c46e6141e3a671cbb9c8054df93e487d78986a320 Nov 24 08:08:39 crc kubenswrapper[4799]: I1124 08:08:39.977901 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:08:40 crc kubenswrapper[4799]: I1124 08:08:40.549360 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-78cxl" event={"ID":"6698e591-c23a-477e-890e-cddb65f66fad","Type":"ContainerStarted","Data":"b17c470d7411aafea8ddce8c46e6141e3a671cbb9c8054df93e487d78986a320"} Nov 24 08:08:40 crc kubenswrapper[4799]: I1124 08:08:40.551203 4799 generic.go:334] "Generic (PLEG): container finished" podID="3f8cae7e-b65e-4104-86b8-f2c6fa283f35" containerID="ef6a9d756bb3f1f53d16a4a16d620f93f94c1d614e5c9efa62b3e82697dc798f" exitCode=0 Nov 24 08:08:40 crc kubenswrapper[4799]: I1124 08:08:40.551255 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" event={"ID":"3f8cae7e-b65e-4104-86b8-f2c6fa283f35","Type":"ContainerDied","Data":"ef6a9d756bb3f1f53d16a4a16d620f93f94c1d614e5c9efa62b3e82697dc798f"} Nov 24 08:08:40 crc kubenswrapper[4799]: I1124 08:08:40.551287 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" event={"ID":"3f8cae7e-b65e-4104-86b8-f2c6fa283f35","Type":"ContainerStarted","Data":"8ba29457ddfa81ff4b8f69ee286bd5612b88ba4b2b1a632bf3575978aa45483e"} Nov 24 08:08:41 crc kubenswrapper[4799]: I1124 08:08:41.565092 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" event={"ID":"3f8cae7e-b65e-4104-86b8-f2c6fa283f35","Type":"ContainerStarted","Data":"d16ac7e0747a57083af24878e129a3b8295cdef68f6df2dea9e2c771a62c8095"} Nov 24 08:08:41 crc kubenswrapper[4799]: I1124 08:08:41.565587 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:41 crc kubenswrapper[4799]: I1124 08:08:41.593860 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" podStartSLOduration=3.59382368 podStartE2EDuration="3.59382368s" podCreationTimestamp="2025-11-24 08:08:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:08:41.586728833 +0000 UTC m=+4867.242711317" watchObservedRunningTime="2025-11-24 08:08:41.59382368 +0000 UTC m=+4867.249806154" Nov 24 08:08:44 crc kubenswrapper[4799]: I1124 08:08:44.593788 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-78cxl" event={"ID":"6698e591-c23a-477e-890e-cddb65f66fad","Type":"ContainerStarted","Data":"35da31ae2944ee1471c48bdc7ba0ebee08a2b4dcaeefc303180628ae20ab14fb"} Nov 24 08:08:44 crc kubenswrapper[4799]: I1124 08:08:44.619422 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-78cxl" podStartSLOduration=2.441320628 podStartE2EDuration="6.619397524s" podCreationTimestamp="2025-11-24 08:08:38 +0000 UTC" firstStartedPulling="2025-11-24 08:08:39.977631115 +0000 UTC m=+4865.633613589" lastFinishedPulling="2025-11-24 08:08:44.155708011 +0000 UTC m=+4869.811690485" observedRunningTime="2025-11-24 08:08:44.610167567 +0000 UTC m=+4870.266150051" watchObservedRunningTime="2025-11-24 08:08:44.619397524 +0000 UTC m=+4870.275380018" Nov 24 08:08:46 crc kubenswrapper[4799]: I1124 08:08:46.613398 4799 generic.go:334] "Generic (PLEG): container finished" podID="6698e591-c23a-477e-890e-cddb65f66fad" containerID="35da31ae2944ee1471c48bdc7ba0ebee08a2b4dcaeefc303180628ae20ab14fb" exitCode=0 Nov 24 08:08:46 crc kubenswrapper[4799]: I1124 08:08:46.613609 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-78cxl" event={"ID":"6698e591-c23a-477e-890e-cddb65f66fad","Type":"ContainerDied","Data":"35da31ae2944ee1471c48bdc7ba0ebee08a2b4dcaeefc303180628ae20ab14fb"} Nov 24 08:08:47 crc kubenswrapper[4799]: I1124 08:08:47.995710 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.092702 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-scripts\") pod \"6698e591-c23a-477e-890e-cddb65f66fad\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.093216 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-combined-ca-bundle\") pod \"6698e591-c23a-477e-890e-cddb65f66fad\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.093320 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6698e591-c23a-477e-890e-cddb65f66fad-logs\") pod \"6698e591-c23a-477e-890e-cddb65f66fad\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.093483 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m2cv\" (UniqueName: \"kubernetes.io/projected/6698e591-c23a-477e-890e-cddb65f66fad-kube-api-access-2m2cv\") pod \"6698e591-c23a-477e-890e-cddb65f66fad\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.093644 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-config-data\") pod \"6698e591-c23a-477e-890e-cddb65f66fad\" (UID: \"6698e591-c23a-477e-890e-cddb65f66fad\") " Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.093793 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6698e591-c23a-477e-890e-cddb65f66fad-logs" (OuterVolumeSpecName: "logs") pod "6698e591-c23a-477e-890e-cddb65f66fad" (UID: "6698e591-c23a-477e-890e-cddb65f66fad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.094166 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6698e591-c23a-477e-890e-cddb65f66fad-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.099085 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-scripts" (OuterVolumeSpecName: "scripts") pod "6698e591-c23a-477e-890e-cddb65f66fad" (UID: "6698e591-c23a-477e-890e-cddb65f66fad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.099290 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6698e591-c23a-477e-890e-cddb65f66fad-kube-api-access-2m2cv" (OuterVolumeSpecName: "kube-api-access-2m2cv") pod "6698e591-c23a-477e-890e-cddb65f66fad" (UID: "6698e591-c23a-477e-890e-cddb65f66fad"). InnerVolumeSpecName "kube-api-access-2m2cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.117374 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6698e591-c23a-477e-890e-cddb65f66fad" (UID: "6698e591-c23a-477e-890e-cddb65f66fad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.121013 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-config-data" (OuterVolumeSpecName: "config-data") pod "6698e591-c23a-477e-890e-cddb65f66fad" (UID: "6698e591-c23a-477e-890e-cddb65f66fad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.196627 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.196661 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.196701 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6698e591-c23a-477e-890e-cddb65f66fad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.196712 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m2cv\" (UniqueName: \"kubernetes.io/projected/6698e591-c23a-477e-890e-cddb65f66fad-kube-api-access-2m2cv\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.632103 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-78cxl" event={"ID":"6698e591-c23a-477e-890e-cddb65f66fad","Type":"ContainerDied","Data":"b17c470d7411aafea8ddce8c46e6141e3a671cbb9c8054df93e487d78986a320"} Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.632145 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b17c470d7411aafea8ddce8c46e6141e3a671cbb9c8054df93e487d78986a320" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.632197 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-78cxl" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.711342 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-bb89858b4-prpnr"] Nov 24 08:08:48 crc kubenswrapper[4799]: E1124 08:08:48.711687 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6698e591-c23a-477e-890e-cddb65f66fad" containerName="placement-db-sync" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.711703 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6698e591-c23a-477e-890e-cddb65f66fad" containerName="placement-db-sync" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.711943 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6698e591-c23a-477e-890e-cddb65f66fad" containerName="placement-db-sync" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.712868 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.715791 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.715828 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-p8cg9" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.716191 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.748743 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bb89858b4-prpnr"] Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.806253 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-logs\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.806320 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-combined-ca-bundle\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.806471 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-config-data\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.806558 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7mrb\" (UniqueName: \"kubernetes.io/projected/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-kube-api-access-v7mrb\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.806669 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-scripts\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.908141 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-config-data\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.908240 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7mrb\" (UniqueName: \"kubernetes.io/projected/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-kube-api-access-v7mrb\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.908296 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-scripts\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.908332 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-logs\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.908355 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-combined-ca-bundle\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.909583 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-logs\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.912468 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-config-data\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.913167 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-scripts\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.917437 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-combined-ca-bundle\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:48 crc kubenswrapper[4799]: I1124 08:08:48.936949 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7mrb\" (UniqueName: \"kubernetes.io/projected/6c4ede8a-4779-4b8e-87c2-9d54664e3f4f-kube-api-access-v7mrb\") pod \"placement-bb89858b4-prpnr\" (UID: \"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f\") " pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.041522 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.127063 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.209972 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6664485f69-r9nj4"] Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.210557 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" podUID="3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" containerName="dnsmasq-dns" containerID="cri-o://cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4" gracePeriod=10 Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.559538 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bb89858b4-prpnr"] Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.648776 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.654237 4799 generic.go:334] "Generic (PLEG): container finished" podID="3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" containerID="cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4" exitCode=0 Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.654405 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.654883 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" event={"ID":"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532","Type":"ContainerDied","Data":"cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4"} Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.654927 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6664485f69-r9nj4" event={"ID":"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532","Type":"ContainerDied","Data":"44168d28414276d8ef58c7b1210b16c06e585fb50cfb430d96b4a2d0670f1da9"} Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.654948 4799 scope.go:117] "RemoveContainer" containerID="cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.658147 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bb89858b4-prpnr" event={"ID":"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f","Type":"ContainerStarted","Data":"74d3c8cde7587daca5bb702b0c0cbe6a3a4963df60c1f09bdb6a614356c08491"} Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.686511 4799 scope.go:117] "RemoveContainer" containerID="bc3c6afbfe50f85305b930f45daae4a2e69cc4ab992b620b03793e52bffefb56" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.725134 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pfd4\" (UniqueName: \"kubernetes.io/projected/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-kube-api-access-9pfd4\") pod \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.725272 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-config\") pod \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.725426 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-dns-svc\") pod \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.725452 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-nb\") pod \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.725510 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-sb\") pod \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\" (UID: \"3e8fc5fa-8dbe-4ce3-9c96-53e58164a532\") " Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.728333 4799 scope.go:117] "RemoveContainer" containerID="cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4" Nov 24 08:08:49 crc kubenswrapper[4799]: E1124 08:08:49.729329 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4\": container with ID starting with cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4 not found: ID does not exist" containerID="cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.729370 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4"} err="failed to get container status \"cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4\": rpc error: code = NotFound desc = could not find container \"cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4\": container with ID starting with cc8882081b4166d6f7d53a97d3652c639e3acaf9513f0bfe7ca3dfcb0f3e37a4 not found: ID does not exist" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.729397 4799 scope.go:117] "RemoveContainer" containerID="bc3c6afbfe50f85305b930f45daae4a2e69cc4ab992b620b03793e52bffefb56" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.729473 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-kube-api-access-9pfd4" (OuterVolumeSpecName: "kube-api-access-9pfd4") pod "3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" (UID: "3e8fc5fa-8dbe-4ce3-9c96-53e58164a532"). InnerVolumeSpecName "kube-api-access-9pfd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:08:49 crc kubenswrapper[4799]: E1124 08:08:49.729808 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc3c6afbfe50f85305b930f45daae4a2e69cc4ab992b620b03793e52bffefb56\": container with ID starting with bc3c6afbfe50f85305b930f45daae4a2e69cc4ab992b620b03793e52bffefb56 not found: ID does not exist" containerID="bc3c6afbfe50f85305b930f45daae4a2e69cc4ab992b620b03793e52bffefb56" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.729893 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc3c6afbfe50f85305b930f45daae4a2e69cc4ab992b620b03793e52bffefb56"} err="failed to get container status \"bc3c6afbfe50f85305b930f45daae4a2e69cc4ab992b620b03793e52bffefb56\": rpc error: code = NotFound desc = could not find container \"bc3c6afbfe50f85305b930f45daae4a2e69cc4ab992b620b03793e52bffefb56\": container with ID starting with bc3c6afbfe50f85305b930f45daae4a2e69cc4ab992b620b03793e52bffefb56 not found: ID does not exist" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.768022 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" (UID: "3e8fc5fa-8dbe-4ce3-9c96-53e58164a532"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.770943 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" (UID: "3e8fc5fa-8dbe-4ce3-9c96-53e58164a532"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.777662 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" (UID: "3e8fc5fa-8dbe-4ce3-9c96-53e58164a532"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.777928 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-config" (OuterVolumeSpecName: "config") pod "3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" (UID: "3e8fc5fa-8dbe-4ce3-9c96-53e58164a532"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.828213 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.828247 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.828257 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.828265 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pfd4\" (UniqueName: \"kubernetes.io/projected/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-kube-api-access-9pfd4\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.828276 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.989459 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6664485f69-r9nj4"] Nov 24 08:08:49 crc kubenswrapper[4799]: I1124 08:08:49.995767 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6664485f69-r9nj4"] Nov 24 08:08:50 crc kubenswrapper[4799]: I1124 08:08:50.675631 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bb89858b4-prpnr" event={"ID":"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f","Type":"ContainerStarted","Data":"5219857736f9f9dc30f8ed21c46a4dceef9059993808fc1e34c63aa693af67fe"} Nov 24 08:08:50 crc kubenswrapper[4799]: I1124 08:08:50.675682 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bb89858b4-prpnr" event={"ID":"6c4ede8a-4779-4b8e-87c2-9d54664e3f4f","Type":"ContainerStarted","Data":"1fbd90c984b13df42ec8c28a51418f1d0ab1cdfc9142efa1535958447399d12a"} Nov 24 08:08:50 crc kubenswrapper[4799]: I1124 08:08:50.675988 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:50 crc kubenswrapper[4799]: I1124 08:08:50.698790 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-bb89858b4-prpnr" podStartSLOduration=2.6987686870000003 podStartE2EDuration="2.698768687s" podCreationTimestamp="2025-11-24 08:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:08:50.696890394 +0000 UTC m=+4876.352872938" watchObservedRunningTime="2025-11-24 08:08:50.698768687 +0000 UTC m=+4876.354751161" Nov 24 08:08:51 crc kubenswrapper[4799]: I1124 08:08:51.650071 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" path="/var/lib/kubelet/pods/3e8fc5fa-8dbe-4ce3-9c96-53e58164a532/volumes" Nov 24 08:08:51 crc kubenswrapper[4799]: I1124 08:08:51.686906 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:08:58 crc kubenswrapper[4799]: I1124 08:08:58.919506 4799 scope.go:117] "RemoveContainer" containerID="a17fa8cb688753166d5579b8d68314f312d32ea8181ffd85e54a8a3fdf260df5" Nov 24 08:08:58 crc kubenswrapper[4799]: I1124 08:08:58.953893 4799 scope.go:117] "RemoveContainer" containerID="36af58232f124ebd95ee1c8c030dcf54d41894292ac10278c731360025a59ed8" Nov 24 08:09:20 crc kubenswrapper[4799]: I1124 08:09:20.026107 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:09:20 crc kubenswrapper[4799]: I1124 08:09:20.034768 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-bb89858b4-prpnr" Nov 24 08:09:20 crc kubenswrapper[4799]: I1124 08:09:20.400277 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:09:20 crc kubenswrapper[4799]: I1124 08:09:20.400329 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.377661 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-cqc6t"] Nov 24 08:09:41 crc kubenswrapper[4799]: E1124 08:09:41.378530 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" containerName="init" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.378545 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" containerName="init" Nov 24 08:09:41 crc kubenswrapper[4799]: E1124 08:09:41.378578 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" containerName="dnsmasq-dns" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.378584 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" containerName="dnsmasq-dns" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.378725 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e8fc5fa-8dbe-4ce3-9c96-53e58164a532" containerName="dnsmasq-dns" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.379313 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cqc6t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.394887 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-cqc6t"] Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.476319 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-ms9fw"] Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.478081 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ms9fw" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.490404 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0026133-9917-4f5c-a465-d55de6765574-operator-scripts\") pod \"nova-api-db-create-cqc6t\" (UID: \"f0026133-9917-4f5c-a465-d55de6765574\") " pod="openstack/nova-api-db-create-cqc6t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.490511 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p5t7\" (UniqueName: \"kubernetes.io/projected/f0026133-9917-4f5c-a465-d55de6765574-kube-api-access-9p5t7\") pod \"nova-api-db-create-cqc6t\" (UID: \"f0026133-9917-4f5c-a465-d55de6765574\") " pod="openstack/nova-api-db-create-cqc6t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.499875 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-ms9fw"] Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.579533 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-846e-account-create-x6p7t"] Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.581097 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-846e-account-create-x6p7t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.582831 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.591575 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0026133-9917-4f5c-a465-d55de6765574-operator-scripts\") pod \"nova-api-db-create-cqc6t\" (UID: \"f0026133-9917-4f5c-a465-d55de6765574\") " pod="openstack/nova-api-db-create-cqc6t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.591836 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-945ph\" (UniqueName: \"kubernetes.io/projected/16703067-39c9-444d-a8d3-f17344f8fe3c-kube-api-access-945ph\") pod \"nova-cell0-db-create-ms9fw\" (UID: \"16703067-39c9-444d-a8d3-f17344f8fe3c\") " pod="openstack/nova-cell0-db-create-ms9fw" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.591997 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p5t7\" (UniqueName: \"kubernetes.io/projected/f0026133-9917-4f5c-a465-d55de6765574-kube-api-access-9p5t7\") pod \"nova-api-db-create-cqc6t\" (UID: \"f0026133-9917-4f5c-a465-d55de6765574\") " pod="openstack/nova-api-db-create-cqc6t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.592159 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16703067-39c9-444d-a8d3-f17344f8fe3c-operator-scripts\") pod \"nova-cell0-db-create-ms9fw\" (UID: \"16703067-39c9-444d-a8d3-f17344f8fe3c\") " pod="openstack/nova-cell0-db-create-ms9fw" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.592973 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0026133-9917-4f5c-a465-d55de6765574-operator-scripts\") pod \"nova-api-db-create-cqc6t\" (UID: \"f0026133-9917-4f5c-a465-d55de6765574\") " pod="openstack/nova-api-db-create-cqc6t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.600225 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-846e-account-create-x6p7t"] Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.679624 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-cjhsm"] Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.680737 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cjhsm" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.687408 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-cjhsm"] Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.693720 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-945ph\" (UniqueName: \"kubernetes.io/projected/16703067-39c9-444d-a8d3-f17344f8fe3c-kube-api-access-945ph\") pod \"nova-cell0-db-create-ms9fw\" (UID: \"16703067-39c9-444d-a8d3-f17344f8fe3c\") " pod="openstack/nova-cell0-db-create-ms9fw" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.693752 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aacd0722-6f99-4a60-a497-a76e96a5f27e-operator-scripts\") pod \"nova-api-846e-account-create-x6p7t\" (UID: \"aacd0722-6f99-4a60-a497-a76e96a5f27e\") " pod="openstack/nova-api-846e-account-create-x6p7t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.693880 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsrv2\" (UniqueName: \"kubernetes.io/projected/aacd0722-6f99-4a60-a497-a76e96a5f27e-kube-api-access-bsrv2\") pod \"nova-api-846e-account-create-x6p7t\" (UID: \"aacd0722-6f99-4a60-a497-a76e96a5f27e\") " pod="openstack/nova-api-846e-account-create-x6p7t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.693899 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16703067-39c9-444d-a8d3-f17344f8fe3c-operator-scripts\") pod \"nova-cell0-db-create-ms9fw\" (UID: \"16703067-39c9-444d-a8d3-f17344f8fe3c\") " pod="openstack/nova-cell0-db-create-ms9fw" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.695423 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16703067-39c9-444d-a8d3-f17344f8fe3c-operator-scripts\") pod \"nova-cell0-db-create-ms9fw\" (UID: \"16703067-39c9-444d-a8d3-f17344f8fe3c\") " pod="openstack/nova-cell0-db-create-ms9fw" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.773027 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p5t7\" (UniqueName: \"kubernetes.io/projected/f0026133-9917-4f5c-a465-d55de6765574-kube-api-access-9p5t7\") pod \"nova-api-db-create-cqc6t\" (UID: \"f0026133-9917-4f5c-a465-d55de6765574\") " pod="openstack/nova-api-db-create-cqc6t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.783814 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-945ph\" (UniqueName: \"kubernetes.io/projected/16703067-39c9-444d-a8d3-f17344f8fe3c-kube-api-access-945ph\") pod \"nova-cell0-db-create-ms9fw\" (UID: \"16703067-39c9-444d-a8d3-f17344f8fe3c\") " pod="openstack/nova-cell0-db-create-ms9fw" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.792126 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-45f0-account-create-q4p2p"] Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.794543 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-45f0-account-create-q4p2p" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.796009 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d13f370-86ba-485c-9108-77b4149c9d3f-operator-scripts\") pod \"nova-cell1-db-create-cjhsm\" (UID: \"2d13f370-86ba-485c-9108-77b4149c9d3f\") " pod="openstack/nova-cell1-db-create-cjhsm" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.796216 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsrv2\" (UniqueName: \"kubernetes.io/projected/aacd0722-6f99-4a60-a497-a76e96a5f27e-kube-api-access-bsrv2\") pod \"nova-api-846e-account-create-x6p7t\" (UID: \"aacd0722-6f99-4a60-a497-a76e96a5f27e\") " pod="openstack/nova-api-846e-account-create-x6p7t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.796345 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6cwb\" (UniqueName: \"kubernetes.io/projected/2d13f370-86ba-485c-9108-77b4149c9d3f-kube-api-access-t6cwb\") pod \"nova-cell1-db-create-cjhsm\" (UID: \"2d13f370-86ba-485c-9108-77b4149c9d3f\") " pod="openstack/nova-cell1-db-create-cjhsm" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.796370 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aacd0722-6f99-4a60-a497-a76e96a5f27e-operator-scripts\") pod \"nova-api-846e-account-create-x6p7t\" (UID: \"aacd0722-6f99-4a60-a497-a76e96a5f27e\") " pod="openstack/nova-api-846e-account-create-x6p7t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.797037 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aacd0722-6f99-4a60-a497-a76e96a5f27e-operator-scripts\") pod \"nova-api-846e-account-create-x6p7t\" (UID: \"aacd0722-6f99-4a60-a497-a76e96a5f27e\") " pod="openstack/nova-api-846e-account-create-x6p7t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.801292 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.802131 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ms9fw" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.816543 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsrv2\" (UniqueName: \"kubernetes.io/projected/aacd0722-6f99-4a60-a497-a76e96a5f27e-kube-api-access-bsrv2\") pod \"nova-api-846e-account-create-x6p7t\" (UID: \"aacd0722-6f99-4a60-a497-a76e96a5f27e\") " pod="openstack/nova-api-846e-account-create-x6p7t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.821833 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-45f0-account-create-q4p2p"] Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.901349 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6cwb\" (UniqueName: \"kubernetes.io/projected/2d13f370-86ba-485c-9108-77b4149c9d3f-kube-api-access-t6cwb\") pod \"nova-cell1-db-create-cjhsm\" (UID: \"2d13f370-86ba-485c-9108-77b4149c9d3f\") " pod="openstack/nova-cell1-db-create-cjhsm" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.902291 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-operator-scripts\") pod \"nova-cell0-45f0-account-create-q4p2p\" (UID: \"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c\") " pod="openstack/nova-cell0-45f0-account-create-q4p2p" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.902391 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d13f370-86ba-485c-9108-77b4149c9d3f-operator-scripts\") pod \"nova-cell1-db-create-cjhsm\" (UID: \"2d13f370-86ba-485c-9108-77b4149c9d3f\") " pod="openstack/nova-cell1-db-create-cjhsm" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.902458 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8dt2\" (UniqueName: \"kubernetes.io/projected/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-kube-api-access-r8dt2\") pod \"nova-cell0-45f0-account-create-q4p2p\" (UID: \"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c\") " pod="openstack/nova-cell0-45f0-account-create-q4p2p" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.903195 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d13f370-86ba-485c-9108-77b4149c9d3f-operator-scripts\") pod \"nova-cell1-db-create-cjhsm\" (UID: \"2d13f370-86ba-485c-9108-77b4149c9d3f\") " pod="openstack/nova-cell1-db-create-cjhsm" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.903708 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-846e-account-create-x6p7t" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.921954 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6cwb\" (UniqueName: \"kubernetes.io/projected/2d13f370-86ba-485c-9108-77b4149c9d3f-kube-api-access-t6cwb\") pod \"nova-cell1-db-create-cjhsm\" (UID: \"2d13f370-86ba-485c-9108-77b4149c9d3f\") " pod="openstack/nova-cell1-db-create-cjhsm" Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.992144 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-8bfc-account-create-vvn6w"] Nov 24 08:09:41 crc kubenswrapper[4799]: I1124 08:09:41.993522 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bfc-account-create-vvn6w" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:41.998752 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.000030 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cjhsm" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.004076 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-operator-scripts\") pod \"nova-cell0-45f0-account-create-q4p2p\" (UID: \"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c\") " pod="openstack/nova-cell0-45f0-account-create-q4p2p" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.004150 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8dt2\" (UniqueName: \"kubernetes.io/projected/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-kube-api-access-r8dt2\") pod \"nova-cell0-45f0-account-create-q4p2p\" (UID: \"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c\") " pod="openstack/nova-cell0-45f0-account-create-q4p2p" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.004804 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-operator-scripts\") pod \"nova-cell0-45f0-account-create-q4p2p\" (UID: \"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c\") " pod="openstack/nova-cell0-45f0-account-create-q4p2p" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.008032 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8bfc-account-create-vvn6w"] Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.017238 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cqc6t" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.029023 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8dt2\" (UniqueName: \"kubernetes.io/projected/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-kube-api-access-r8dt2\") pod \"nova-cell0-45f0-account-create-q4p2p\" (UID: \"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c\") " pod="openstack/nova-cell0-45f0-account-create-q4p2p" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.105683 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg24h\" (UniqueName: \"kubernetes.io/projected/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-kube-api-access-jg24h\") pod \"nova-cell1-8bfc-account-create-vvn6w\" (UID: \"a18b4911-5cfb-4385-a5c5-8f7bf37605d8\") " pod="openstack/nova-cell1-8bfc-account-create-vvn6w" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.106068 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-operator-scripts\") pod \"nova-cell1-8bfc-account-create-vvn6w\" (UID: \"a18b4911-5cfb-4385-a5c5-8f7bf37605d8\") " pod="openstack/nova-cell1-8bfc-account-create-vvn6w" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.113633 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-45f0-account-create-q4p2p" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.207487 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg24h\" (UniqueName: \"kubernetes.io/projected/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-kube-api-access-jg24h\") pod \"nova-cell1-8bfc-account-create-vvn6w\" (UID: \"a18b4911-5cfb-4385-a5c5-8f7bf37605d8\") " pod="openstack/nova-cell1-8bfc-account-create-vvn6w" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.207640 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-operator-scripts\") pod \"nova-cell1-8bfc-account-create-vvn6w\" (UID: \"a18b4911-5cfb-4385-a5c5-8f7bf37605d8\") " pod="openstack/nova-cell1-8bfc-account-create-vvn6w" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.208961 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-operator-scripts\") pod \"nova-cell1-8bfc-account-create-vvn6w\" (UID: \"a18b4911-5cfb-4385-a5c5-8f7bf37605d8\") " pod="openstack/nova-cell1-8bfc-account-create-vvn6w" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.224565 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg24h\" (UniqueName: \"kubernetes.io/projected/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-kube-api-access-jg24h\") pod \"nova-cell1-8bfc-account-create-vvn6w\" (UID: \"a18b4911-5cfb-4385-a5c5-8f7bf37605d8\") " pod="openstack/nova-cell1-8bfc-account-create-vvn6w" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.297680 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-ms9fw"] Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.424807 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bfc-account-create-vvn6w" Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.446803 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-846e-account-create-x6p7t"] Nov 24 08:09:42 crc kubenswrapper[4799]: W1124 08:09:42.450936 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaacd0722_6f99_4a60_a497_a76e96a5f27e.slice/crio-e48263c65a5f3fce96062d48731f91643c9ebbf85b27c18411e7e65fce827365 WatchSource:0}: Error finding container e48263c65a5f3fce96062d48731f91643c9ebbf85b27c18411e7e65fce827365: Status 404 returned error can't find the container with id e48263c65a5f3fce96062d48731f91643c9ebbf85b27c18411e7e65fce827365 Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.646649 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-cqc6t"] Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.755694 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-cjhsm"] Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.848376 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-45f0-account-create-q4p2p"] Nov 24 08:09:42 crc kubenswrapper[4799]: W1124 08:09:42.876073 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0c0c72c_85ae_4dca_8c60_58a9bef04b3c.slice/crio-409650b902b1f4e25ae05ef8d43079607db5379071f32c708162ed2f82a98ccd WatchSource:0}: Error finding container 409650b902b1f4e25ae05ef8d43079607db5379071f32c708162ed2f82a98ccd: Status 404 returned error can't find the container with id 409650b902b1f4e25ae05ef8d43079607db5379071f32c708162ed2f82a98ccd Nov 24 08:09:42 crc kubenswrapper[4799]: I1124 08:09:42.967695 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8bfc-account-create-vvn6w"] Nov 24 08:09:43 crc kubenswrapper[4799]: W1124 08:09:43.027152 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda18b4911_5cfb_4385_a5c5_8f7bf37605d8.slice/crio-7187fe6d366545cce33aa86f583f25dc6079810cbba8c73dd79143febc1a287d WatchSource:0}: Error finding container 7187fe6d366545cce33aa86f583f25dc6079810cbba8c73dd79143febc1a287d: Status 404 returned error can't find the container with id 7187fe6d366545cce33aa86f583f25dc6079810cbba8c73dd79143febc1a287d Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.169644 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-cjhsm" event={"ID":"2d13f370-86ba-485c-9108-77b4149c9d3f","Type":"ContainerStarted","Data":"8bdea0a0a8ef6b0859843eb10b4e56bced7bcb2b4d2271efb91c2d691cfb8c99"} Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.170876 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8bfc-account-create-vvn6w" event={"ID":"a18b4911-5cfb-4385-a5c5-8f7bf37605d8","Type":"ContainerStarted","Data":"7187fe6d366545cce33aa86f583f25dc6079810cbba8c73dd79143febc1a287d"} Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.171707 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-45f0-account-create-q4p2p" event={"ID":"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c","Type":"ContainerStarted","Data":"409650b902b1f4e25ae05ef8d43079607db5379071f32c708162ed2f82a98ccd"} Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.173472 4799 generic.go:334] "Generic (PLEG): container finished" podID="16703067-39c9-444d-a8d3-f17344f8fe3c" containerID="0e2d4a36f9e6eff46d6847b5020133267a5059afebf953b3fbdddd0a74909b92" exitCode=0 Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.173528 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ms9fw" event={"ID":"16703067-39c9-444d-a8d3-f17344f8fe3c","Type":"ContainerDied","Data":"0e2d4a36f9e6eff46d6847b5020133267a5059afebf953b3fbdddd0a74909b92"} Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.173543 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ms9fw" event={"ID":"16703067-39c9-444d-a8d3-f17344f8fe3c","Type":"ContainerStarted","Data":"268499cc17a0363024ca014675af53a6f025e6baed10bd2a50404a773f50fac8"} Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.174964 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cqc6t" event={"ID":"f0026133-9917-4f5c-a465-d55de6765574","Type":"ContainerStarted","Data":"155e92ba0a2b32b54a807ec997bd5b62d4480d793f78e290425e3d0f77993f58"} Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.174994 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cqc6t" event={"ID":"f0026133-9917-4f5c-a465-d55de6765574","Type":"ContainerStarted","Data":"ff51effec2c9beace16751f64d2f2368da23ab77d171d540d200bfb768193883"} Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.177998 4799 generic.go:334] "Generic (PLEG): container finished" podID="aacd0722-6f99-4a60-a497-a76e96a5f27e" containerID="56cdb01ffc7f232fdb2dd2b14aef6f670361800f20ecf6599b58f6ad3c208d63" exitCode=0 Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.178060 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-846e-account-create-x6p7t" event={"ID":"aacd0722-6f99-4a60-a497-a76e96a5f27e","Type":"ContainerDied","Data":"56cdb01ffc7f232fdb2dd2b14aef6f670361800f20ecf6599b58f6ad3c208d63"} Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.178091 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-846e-account-create-x6p7t" event={"ID":"aacd0722-6f99-4a60-a497-a76e96a5f27e","Type":"ContainerStarted","Data":"e48263c65a5f3fce96062d48731f91643c9ebbf85b27c18411e7e65fce827365"} Nov 24 08:09:43 crc kubenswrapper[4799]: I1124 08:09:43.211498 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-cqc6t" podStartSLOduration=2.211471268 podStartE2EDuration="2.211471268s" podCreationTimestamp="2025-11-24 08:09:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:09:43.20469991 +0000 UTC m=+4928.860682394" watchObservedRunningTime="2025-11-24 08:09:43.211471268 +0000 UTC m=+4928.867453762" Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.187465 4799 generic.go:334] "Generic (PLEG): container finished" podID="a18b4911-5cfb-4385-a5c5-8f7bf37605d8" containerID="44e5b6902a29ae87dbad1efa26ef3da911a1bef156b846bd59f80b2933fb7d5a" exitCode=0 Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.187717 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8bfc-account-create-vvn6w" event={"ID":"a18b4911-5cfb-4385-a5c5-8f7bf37605d8","Type":"ContainerDied","Data":"44e5b6902a29ae87dbad1efa26ef3da911a1bef156b846bd59f80b2933fb7d5a"} Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.191518 4799 generic.go:334] "Generic (PLEG): container finished" podID="c0c0c72c-85ae-4dca-8c60-58a9bef04b3c" containerID="bb39807688abcd5f157e17009306b743ade6d35ed4c3156f413549594b846226" exitCode=0 Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.191631 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-45f0-account-create-q4p2p" event={"ID":"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c","Type":"ContainerDied","Data":"bb39807688abcd5f157e17009306b743ade6d35ed4c3156f413549594b846226"} Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.193464 4799 generic.go:334] "Generic (PLEG): container finished" podID="f0026133-9917-4f5c-a465-d55de6765574" containerID="155e92ba0a2b32b54a807ec997bd5b62d4480d793f78e290425e3d0f77993f58" exitCode=0 Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.193590 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cqc6t" event={"ID":"f0026133-9917-4f5c-a465-d55de6765574","Type":"ContainerDied","Data":"155e92ba0a2b32b54a807ec997bd5b62d4480d793f78e290425e3d0f77993f58"} Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.194811 4799 generic.go:334] "Generic (PLEG): container finished" podID="2d13f370-86ba-485c-9108-77b4149c9d3f" containerID="598c7ea3ac2465ec38f036baebd45b07a5e38a00486d329f3551ab9f0b85dfb0" exitCode=0 Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.194941 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-cjhsm" event={"ID":"2d13f370-86ba-485c-9108-77b4149c9d3f","Type":"ContainerDied","Data":"598c7ea3ac2465ec38f036baebd45b07a5e38a00486d329f3551ab9f0b85dfb0"} Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.612192 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-846e-account-create-x6p7t" Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.619193 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ms9fw" Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.757184 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsrv2\" (UniqueName: \"kubernetes.io/projected/aacd0722-6f99-4a60-a497-a76e96a5f27e-kube-api-access-bsrv2\") pod \"aacd0722-6f99-4a60-a497-a76e96a5f27e\" (UID: \"aacd0722-6f99-4a60-a497-a76e96a5f27e\") " Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.757308 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aacd0722-6f99-4a60-a497-a76e96a5f27e-operator-scripts\") pod \"aacd0722-6f99-4a60-a497-a76e96a5f27e\" (UID: \"aacd0722-6f99-4a60-a497-a76e96a5f27e\") " Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.758135 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aacd0722-6f99-4a60-a497-a76e96a5f27e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aacd0722-6f99-4a60-a497-a76e96a5f27e" (UID: "aacd0722-6f99-4a60-a497-a76e96a5f27e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.758271 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16703067-39c9-444d-a8d3-f17344f8fe3c-operator-scripts\") pod \"16703067-39c9-444d-a8d3-f17344f8fe3c\" (UID: \"16703067-39c9-444d-a8d3-f17344f8fe3c\") " Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.758379 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-945ph\" (UniqueName: \"kubernetes.io/projected/16703067-39c9-444d-a8d3-f17344f8fe3c-kube-api-access-945ph\") pod \"16703067-39c9-444d-a8d3-f17344f8fe3c\" (UID: \"16703067-39c9-444d-a8d3-f17344f8fe3c\") " Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.758754 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16703067-39c9-444d-a8d3-f17344f8fe3c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16703067-39c9-444d-a8d3-f17344f8fe3c" (UID: "16703067-39c9-444d-a8d3-f17344f8fe3c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.759253 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aacd0722-6f99-4a60-a497-a76e96a5f27e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.759590 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16703067-39c9-444d-a8d3-f17344f8fe3c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.763060 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16703067-39c9-444d-a8d3-f17344f8fe3c-kube-api-access-945ph" (OuterVolumeSpecName: "kube-api-access-945ph") pod "16703067-39c9-444d-a8d3-f17344f8fe3c" (UID: "16703067-39c9-444d-a8d3-f17344f8fe3c"). InnerVolumeSpecName "kube-api-access-945ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.765045 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aacd0722-6f99-4a60-a497-a76e96a5f27e-kube-api-access-bsrv2" (OuterVolumeSpecName: "kube-api-access-bsrv2") pod "aacd0722-6f99-4a60-a497-a76e96a5f27e" (UID: "aacd0722-6f99-4a60-a497-a76e96a5f27e"). InnerVolumeSpecName "kube-api-access-bsrv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.861084 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsrv2\" (UniqueName: \"kubernetes.io/projected/aacd0722-6f99-4a60-a497-a76e96a5f27e-kube-api-access-bsrv2\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:44 crc kubenswrapper[4799]: I1124 08:09:44.861128 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-945ph\" (UniqueName: \"kubernetes.io/projected/16703067-39c9-444d-a8d3-f17344f8fe3c-kube-api-access-945ph\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.215217 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ms9fw" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.215172 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ms9fw" event={"ID":"16703067-39c9-444d-a8d3-f17344f8fe3c","Type":"ContainerDied","Data":"268499cc17a0363024ca014675af53a6f025e6baed10bd2a50404a773f50fac8"} Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.215360 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="268499cc17a0363024ca014675af53a6f025e6baed10bd2a50404a773f50fac8" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.218121 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-846e-account-create-x6p7t" event={"ID":"aacd0722-6f99-4a60-a497-a76e96a5f27e","Type":"ContainerDied","Data":"e48263c65a5f3fce96062d48731f91643c9ebbf85b27c18411e7e65fce827365"} Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.218151 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e48263c65a5f3fce96062d48731f91643c9ebbf85b27c18411e7e65fce827365" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.218590 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-846e-account-create-x6p7t" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.699388 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cjhsm" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.720137 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bfc-account-create-vvn6w" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.727827 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-45f0-account-create-q4p2p" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.731079 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cqc6t" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.784266 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6cwb\" (UniqueName: \"kubernetes.io/projected/2d13f370-86ba-485c-9108-77b4149c9d3f-kube-api-access-t6cwb\") pod \"2d13f370-86ba-485c-9108-77b4149c9d3f\" (UID: \"2d13f370-86ba-485c-9108-77b4149c9d3f\") " Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.784442 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d13f370-86ba-485c-9108-77b4149c9d3f-operator-scripts\") pod \"2d13f370-86ba-485c-9108-77b4149c9d3f\" (UID: \"2d13f370-86ba-485c-9108-77b4149c9d3f\") " Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.787691 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d13f370-86ba-485c-9108-77b4149c9d3f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d13f370-86ba-485c-9108-77b4149c9d3f" (UID: "2d13f370-86ba-485c-9108-77b4149c9d3f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.789171 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d13f370-86ba-485c-9108-77b4149c9d3f-kube-api-access-t6cwb" (OuterVolumeSpecName: "kube-api-access-t6cwb") pod "2d13f370-86ba-485c-9108-77b4149c9d3f" (UID: "2d13f370-86ba-485c-9108-77b4149c9d3f"). InnerVolumeSpecName "kube-api-access-t6cwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.886628 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-operator-scripts\") pod \"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c\" (UID: \"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c\") " Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.886734 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-operator-scripts\") pod \"a18b4911-5cfb-4385-a5c5-8f7bf37605d8\" (UID: \"a18b4911-5cfb-4385-a5c5-8f7bf37605d8\") " Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.886817 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8dt2\" (UniqueName: \"kubernetes.io/projected/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-kube-api-access-r8dt2\") pod \"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c\" (UID: \"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c\") " Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.886837 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg24h\" (UniqueName: \"kubernetes.io/projected/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-kube-api-access-jg24h\") pod \"a18b4911-5cfb-4385-a5c5-8f7bf37605d8\" (UID: \"a18b4911-5cfb-4385-a5c5-8f7bf37605d8\") " Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.886912 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p5t7\" (UniqueName: \"kubernetes.io/projected/f0026133-9917-4f5c-a465-d55de6765574-kube-api-access-9p5t7\") pod \"f0026133-9917-4f5c-a465-d55de6765574\" (UID: \"f0026133-9917-4f5c-a465-d55de6765574\") " Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.886952 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0026133-9917-4f5c-a465-d55de6765574-operator-scripts\") pod \"f0026133-9917-4f5c-a465-d55de6765574\" (UID: \"f0026133-9917-4f5c-a465-d55de6765574\") " Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.887187 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c0c0c72c-85ae-4dca-8c60-58a9bef04b3c" (UID: "c0c0c72c-85ae-4dca-8c60-58a9bef04b3c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.887591 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0026133-9917-4f5c-a465-d55de6765574-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f0026133-9917-4f5c-a465-d55de6765574" (UID: "f0026133-9917-4f5c-a465-d55de6765574"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.887643 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a18b4911-5cfb-4385-a5c5-8f7bf37605d8" (UID: "a18b4911-5cfb-4385-a5c5-8f7bf37605d8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.887668 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d13f370-86ba-485c-9108-77b4149c9d3f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.887683 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.887694 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6cwb\" (UniqueName: \"kubernetes.io/projected/2d13f370-86ba-485c-9108-77b4149c9d3f-kube-api-access-t6cwb\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.890026 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-kube-api-access-r8dt2" (OuterVolumeSpecName: "kube-api-access-r8dt2") pod "c0c0c72c-85ae-4dca-8c60-58a9bef04b3c" (UID: "c0c0c72c-85ae-4dca-8c60-58a9bef04b3c"). InnerVolumeSpecName "kube-api-access-r8dt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.891576 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-kube-api-access-jg24h" (OuterVolumeSpecName: "kube-api-access-jg24h") pod "a18b4911-5cfb-4385-a5c5-8f7bf37605d8" (UID: "a18b4911-5cfb-4385-a5c5-8f7bf37605d8"). InnerVolumeSpecName "kube-api-access-jg24h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.891617 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0026133-9917-4f5c-a465-d55de6765574-kube-api-access-9p5t7" (OuterVolumeSpecName: "kube-api-access-9p5t7") pod "f0026133-9917-4f5c-a465-d55de6765574" (UID: "f0026133-9917-4f5c-a465-d55de6765574"). InnerVolumeSpecName "kube-api-access-9p5t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.989445 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.989479 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg24h\" (UniqueName: \"kubernetes.io/projected/a18b4911-5cfb-4385-a5c5-8f7bf37605d8-kube-api-access-jg24h\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.989494 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8dt2\" (UniqueName: \"kubernetes.io/projected/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c-kube-api-access-r8dt2\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.989507 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p5t7\" (UniqueName: \"kubernetes.io/projected/f0026133-9917-4f5c-a465-d55de6765574-kube-api-access-9p5t7\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:45 crc kubenswrapper[4799]: I1124 08:09:45.989520 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0026133-9917-4f5c-a465-d55de6765574-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.230308 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8bfc-account-create-vvn6w" event={"ID":"a18b4911-5cfb-4385-a5c5-8f7bf37605d8","Type":"ContainerDied","Data":"7187fe6d366545cce33aa86f583f25dc6079810cbba8c73dd79143febc1a287d"} Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.230612 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7187fe6d366545cce33aa86f583f25dc6079810cbba8c73dd79143febc1a287d" Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.230346 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bfc-account-create-vvn6w" Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.234684 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-45f0-account-create-q4p2p" event={"ID":"c0c0c72c-85ae-4dca-8c60-58a9bef04b3c","Type":"ContainerDied","Data":"409650b902b1f4e25ae05ef8d43079607db5379071f32c708162ed2f82a98ccd"} Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.234722 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="409650b902b1f4e25ae05ef8d43079607db5379071f32c708162ed2f82a98ccd" Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.234799 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-45f0-account-create-q4p2p" Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.236617 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cqc6t" event={"ID":"f0026133-9917-4f5c-a465-d55de6765574","Type":"ContainerDied","Data":"ff51effec2c9beace16751f64d2f2368da23ab77d171d540d200bfb768193883"} Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.236641 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff51effec2c9beace16751f64d2f2368da23ab77d171d540d200bfb768193883" Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.236674 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cqc6t" Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.238818 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-cjhsm" event={"ID":"2d13f370-86ba-485c-9108-77b4149c9d3f","Type":"ContainerDied","Data":"8bdea0a0a8ef6b0859843eb10b4e56bced7bcb2b4d2271efb91c2d691cfb8c99"} Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.238855 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cjhsm" Nov 24 08:09:46 crc kubenswrapper[4799]: I1124 08:09:46.238870 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bdea0a0a8ef6b0859843eb10b4e56bced7bcb2b4d2271efb91c2d691cfb8c99" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.057194 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nkdvl"] Nov 24 08:09:47 crc kubenswrapper[4799]: E1124 08:09:47.057600 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aacd0722-6f99-4a60-a497-a76e96a5f27e" containerName="mariadb-account-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.057615 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="aacd0722-6f99-4a60-a497-a76e96a5f27e" containerName="mariadb-account-create" Nov 24 08:09:47 crc kubenswrapper[4799]: E1124 08:09:47.057661 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16703067-39c9-444d-a8d3-f17344f8fe3c" containerName="mariadb-database-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.057670 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="16703067-39c9-444d-a8d3-f17344f8fe3c" containerName="mariadb-database-create" Nov 24 08:09:47 crc kubenswrapper[4799]: E1124 08:09:47.057692 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d13f370-86ba-485c-9108-77b4149c9d3f" containerName="mariadb-database-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.057700 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d13f370-86ba-485c-9108-77b4149c9d3f" containerName="mariadb-database-create" Nov 24 08:09:47 crc kubenswrapper[4799]: E1124 08:09:47.057726 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0026133-9917-4f5c-a465-d55de6765574" containerName="mariadb-database-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.057735 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0026133-9917-4f5c-a465-d55de6765574" containerName="mariadb-database-create" Nov 24 08:09:47 crc kubenswrapper[4799]: E1124 08:09:47.057759 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0c0c72c-85ae-4dca-8c60-58a9bef04b3c" containerName="mariadb-account-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.057766 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0c0c72c-85ae-4dca-8c60-58a9bef04b3c" containerName="mariadb-account-create" Nov 24 08:09:47 crc kubenswrapper[4799]: E1124 08:09:47.057783 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a18b4911-5cfb-4385-a5c5-8f7bf37605d8" containerName="mariadb-account-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.057790 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a18b4911-5cfb-4385-a5c5-8f7bf37605d8" containerName="mariadb-account-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.058035 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a18b4911-5cfb-4385-a5c5-8f7bf37605d8" containerName="mariadb-account-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.058061 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0026133-9917-4f5c-a465-d55de6765574" containerName="mariadb-database-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.058076 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0c0c72c-85ae-4dca-8c60-58a9bef04b3c" containerName="mariadb-account-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.058089 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="16703067-39c9-444d-a8d3-f17344f8fe3c" containerName="mariadb-database-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.058100 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d13f370-86ba-485c-9108-77b4149c9d3f" containerName="mariadb-database-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.058111 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="aacd0722-6f99-4a60-a497-a76e96a5f27e" containerName="mariadb-account-create" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.058808 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.060820 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.061057 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.061190 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-rgc5q" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.075954 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nkdvl"] Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.209152 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-config-data\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.209204 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-scripts\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.209327 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtb4n\" (UniqueName: \"kubernetes.io/projected/57439b00-6d98-498c-bc96-38dfcf395507-kube-api-access-rtb4n\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.209359 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.311210 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-config-data\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.311257 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-scripts\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.311461 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtb4n\" (UniqueName: \"kubernetes.io/projected/57439b00-6d98-498c-bc96-38dfcf395507-kube-api-access-rtb4n\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.311501 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.316157 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-config-data\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.316255 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.317656 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-scripts\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.332240 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtb4n\" (UniqueName: \"kubernetes.io/projected/57439b00-6d98-498c-bc96-38dfcf395507-kube-api-access-rtb4n\") pod \"nova-cell0-conductor-db-sync-nkdvl\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.380709 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:09:47 crc kubenswrapper[4799]: I1124 08:09:47.655773 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nkdvl"] Nov 24 08:09:48 crc kubenswrapper[4799]: I1124 08:09:48.259427 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nkdvl" event={"ID":"57439b00-6d98-498c-bc96-38dfcf395507","Type":"ContainerStarted","Data":"263ea53872fa2b24e9cf5bc0a29f03b64e8a70cb54efb722efb0077225ad33cc"} Nov 24 08:09:50 crc kubenswrapper[4799]: I1124 08:09:50.400804 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:09:50 crc kubenswrapper[4799]: I1124 08:09:50.401291 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:09:58 crc kubenswrapper[4799]: I1124 08:09:58.352367 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nkdvl" event={"ID":"57439b00-6d98-498c-bc96-38dfcf395507","Type":"ContainerStarted","Data":"dbd871a6d8585a23df6a77bdf9df5a46ea1bbca5c517b00db597d57dd0bdb34a"} Nov 24 08:09:58 crc kubenswrapper[4799]: I1124 08:09:58.374745 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-nkdvl" podStartSLOduration=1.757908933 podStartE2EDuration="11.374728091s" podCreationTimestamp="2025-11-24 08:09:47 +0000 UTC" firstStartedPulling="2025-11-24 08:09:47.671767595 +0000 UTC m=+4933.327750069" lastFinishedPulling="2025-11-24 08:09:57.288586743 +0000 UTC m=+4942.944569227" observedRunningTime="2025-11-24 08:09:58.370085792 +0000 UTC m=+4944.026068296" watchObservedRunningTime="2025-11-24 08:09:58.374728091 +0000 UTC m=+4944.030710565" Nov 24 08:09:59 crc kubenswrapper[4799]: I1124 08:09:59.094289 4799 scope.go:117] "RemoveContainer" containerID="5171748c56af678498262fbbe2b2360d5bccdb640b52a24f43c0b0cf2d15493a" Nov 24 08:10:03 crc kubenswrapper[4799]: I1124 08:10:03.411063 4799 generic.go:334] "Generic (PLEG): container finished" podID="57439b00-6d98-498c-bc96-38dfcf395507" containerID="dbd871a6d8585a23df6a77bdf9df5a46ea1bbca5c517b00db597d57dd0bdb34a" exitCode=0 Nov 24 08:10:03 crc kubenswrapper[4799]: I1124 08:10:03.411113 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nkdvl" event={"ID":"57439b00-6d98-498c-bc96-38dfcf395507","Type":"ContainerDied","Data":"dbd871a6d8585a23df6a77bdf9df5a46ea1bbca5c517b00db597d57dd0bdb34a"} Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.748129 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.883972 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-config-data\") pod \"57439b00-6d98-498c-bc96-38dfcf395507\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.884353 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-scripts\") pod \"57439b00-6d98-498c-bc96-38dfcf395507\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.884383 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-combined-ca-bundle\") pod \"57439b00-6d98-498c-bc96-38dfcf395507\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.884461 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtb4n\" (UniqueName: \"kubernetes.io/projected/57439b00-6d98-498c-bc96-38dfcf395507-kube-api-access-rtb4n\") pod \"57439b00-6d98-498c-bc96-38dfcf395507\" (UID: \"57439b00-6d98-498c-bc96-38dfcf395507\") " Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.897040 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57439b00-6d98-498c-bc96-38dfcf395507-kube-api-access-rtb4n" (OuterVolumeSpecName: "kube-api-access-rtb4n") pod "57439b00-6d98-498c-bc96-38dfcf395507" (UID: "57439b00-6d98-498c-bc96-38dfcf395507"). InnerVolumeSpecName "kube-api-access-rtb4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.909507 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-scripts" (OuterVolumeSpecName: "scripts") pod "57439b00-6d98-498c-bc96-38dfcf395507" (UID: "57439b00-6d98-498c-bc96-38dfcf395507"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.913005 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57439b00-6d98-498c-bc96-38dfcf395507" (UID: "57439b00-6d98-498c-bc96-38dfcf395507"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.921643 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-config-data" (OuterVolumeSpecName: "config-data") pod "57439b00-6d98-498c-bc96-38dfcf395507" (UID: "57439b00-6d98-498c-bc96-38dfcf395507"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.987080 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtb4n\" (UniqueName: \"kubernetes.io/projected/57439b00-6d98-498c-bc96-38dfcf395507-kube-api-access-rtb4n\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.987121 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.988959 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:04 crc kubenswrapper[4799]: I1124 08:10:04.989016 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57439b00-6d98-498c-bc96-38dfcf395507-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.444704 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-nkdvl" event={"ID":"57439b00-6d98-498c-bc96-38dfcf395507","Type":"ContainerDied","Data":"263ea53872fa2b24e9cf5bc0a29f03b64e8a70cb54efb722efb0077225ad33cc"} Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.444759 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="263ea53872fa2b24e9cf5bc0a29f03b64e8a70cb54efb722efb0077225ad33cc" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.444835 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-nkdvl" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.511718 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:10:05 crc kubenswrapper[4799]: E1124 08:10:05.512186 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57439b00-6d98-498c-bc96-38dfcf395507" containerName="nova-cell0-conductor-db-sync" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.512203 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="57439b00-6d98-498c-bc96-38dfcf395507" containerName="nova-cell0-conductor-db-sync" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.512379 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="57439b00-6d98-498c-bc96-38dfcf395507" containerName="nova-cell0-conductor-db-sync" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.512978 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.515832 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-rgc5q" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.516002 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.531754 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.700802 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.700909 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.700974 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnxv6\" (UniqueName: \"kubernetes.io/projected/f9749cf2-9a71-453f-88f3-539c15c87c9a-kube-api-access-wnxv6\") pod \"nova-cell0-conductor-0\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.802884 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.803159 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.804831 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnxv6\" (UniqueName: \"kubernetes.io/projected/f9749cf2-9a71-453f-88f3-539c15c87c9a-kube-api-access-wnxv6\") pod \"nova-cell0-conductor-0\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.808959 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.809906 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:05 crc kubenswrapper[4799]: I1124 08:10:05.842631 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnxv6\" (UniqueName: \"kubernetes.io/projected/f9749cf2-9a71-453f-88f3-539c15c87c9a-kube-api-access-wnxv6\") pod \"nova-cell0-conductor-0\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:06 crc kubenswrapper[4799]: I1124 08:10:06.131999 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:06 crc kubenswrapper[4799]: I1124 08:10:06.581582 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:10:07 crc kubenswrapper[4799]: I1124 08:10:07.462516 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f9749cf2-9a71-453f-88f3-539c15c87c9a","Type":"ContainerStarted","Data":"bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f"} Nov 24 08:10:07 crc kubenswrapper[4799]: I1124 08:10:07.462880 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f9749cf2-9a71-453f-88f3-539c15c87c9a","Type":"ContainerStarted","Data":"7051e3f81fa53419831335302245ec1579cf677d3f1d71c63875bac4472033ff"} Nov 24 08:10:07 crc kubenswrapper[4799]: I1124 08:10:07.462900 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:07 crc kubenswrapper[4799]: I1124 08:10:07.476132 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.476111499 podStartE2EDuration="2.476111499s" podCreationTimestamp="2025-11-24 08:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:07.475113311 +0000 UTC m=+4953.131095795" watchObservedRunningTime="2025-11-24 08:10:07.476111499 +0000 UTC m=+4953.132093973" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.166933 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.610600 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-72q6b"] Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.612004 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.616600 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.618436 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.647282 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-72q6b"] Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.724756 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-scripts\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.724817 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-config-data\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.724883 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.724966 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2kmf\" (UniqueName: \"kubernetes.io/projected/d1dc5c43-814a-4a95-91b2-75880697ab67-kube-api-access-z2kmf\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.751614 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.752726 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.762099 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.781548 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.811684 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.812887 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.817520 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.822124 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.826184 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2kmf\" (UniqueName: \"kubernetes.io/projected/d1dc5c43-814a-4a95-91b2-75880697ab67-kube-api-access-z2kmf\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.826241 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-scripts\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.826281 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-config-data\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.826312 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.834782 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.835451 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-config-data\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.845592 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2kmf\" (UniqueName: \"kubernetes.io/projected/d1dc5c43-814a-4a95-91b2-75880697ab67-kube-api-access-z2kmf\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.863534 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-scripts\") pod \"nova-cell0-cell-mapping-72q6b\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.892183 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.894080 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.932221 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.932964 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.933085 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-config-data\") pod \"nova-scheduler-0\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.933114 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m5m2\" (UniqueName: \"kubernetes.io/projected/8e9c42d8-a576-460f-8629-2536bd6efeba-kube-api-access-9m5m2\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.933160 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.933177 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.933219 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bpck\" (UniqueName: \"kubernetes.io/projected/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-kube-api-access-4bpck\") pod \"nova-scheduler-0\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.957201 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:11 crc kubenswrapper[4799]: I1124 08:10:11.990576 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.037057 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bpck\" (UniqueName: \"kubernetes.io/projected/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-kube-api-access-4bpck\") pod \"nova-scheduler-0\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.037134 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.037231 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-config-data\") pod \"nova-scheduler-0\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.037268 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-config-data\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.037296 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m5m2\" (UniqueName: \"kubernetes.io/projected/8e9c42d8-a576-460f-8629-2536bd6efeba-kube-api-access-9m5m2\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.037319 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-logs\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.037343 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.037399 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wp6p\" (UniqueName: \"kubernetes.io/projected/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-kube-api-access-2wp6p\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.037432 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.037459 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.051055 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.062768 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.067573 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.074748 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-config-data\") pod \"nova-scheduler-0\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.081362 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.082776 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.097477 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.109767 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cb57d6b9-bmvcf"] Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.117133 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.118677 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bpck\" (UniqueName: \"kubernetes.io/projected/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-kube-api-access-4bpck\") pod \"nova-scheduler-0\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.131779 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m5m2\" (UniqueName: \"kubernetes.io/projected/8e9c42d8-a576-460f-8629-2536bd6efeba-kube-api-access-9m5m2\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.138991 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-config-data\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.139039 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-logs\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.139063 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.139105 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wp6p\" (UniqueName: \"kubernetes.io/projected/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-kube-api-access-2wp6p\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.148928 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-logs\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.150248 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.153705 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-config-data\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.160712 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.174506 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wp6p\" (UniqueName: \"kubernetes.io/projected/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-kube-api-access-2wp6p\") pod \"nova-metadata-0\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.240482 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cb57d6b9-bmvcf"] Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.245126 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-dns-svc\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.245172 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-config-data\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.245202 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-sb\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.245221 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67e6baf3-e623-4860-9c7f-46cdd5d0d951-logs\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.245241 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc4cf\" (UniqueName: \"kubernetes.io/projected/67e6baf3-e623-4860-9c7f-46cdd5d0d951-kube-api-access-wc4cf\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.245268 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-config\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.245292 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zktlg\" (UniqueName: \"kubernetes.io/projected/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-kube-api-access-zktlg\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.245325 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.245379 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-nb\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.259010 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.278292 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.350177 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.350292 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-nb\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.350348 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-dns-svc\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.350383 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-config-data\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.350419 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-sb\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.350440 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67e6baf3-e623-4860-9c7f-46cdd5d0d951-logs\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.350471 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc4cf\" (UniqueName: \"kubernetes.io/projected/67e6baf3-e623-4860-9c7f-46cdd5d0d951-kube-api-access-wc4cf\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.350507 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-config\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.350546 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zktlg\" (UniqueName: \"kubernetes.io/projected/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-kube-api-access-zktlg\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.352283 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-sb\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.353035 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-nb\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.353428 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67e6baf3-e623-4860-9c7f-46cdd5d0d951-logs\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.354085 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-config\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.355761 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-dns-svc\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.357763 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.371327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-config-data\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.375338 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zktlg\" (UniqueName: \"kubernetes.io/projected/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-kube-api-access-zktlg\") pod \"dnsmasq-dns-59cb57d6b9-bmvcf\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.386052 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.396093 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc4cf\" (UniqueName: \"kubernetes.io/projected/67e6baf3-e623-4860-9c7f-46cdd5d0d951-kube-api-access-wc4cf\") pod \"nova-api-0\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.483414 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.519073 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.795868 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-72q6b"] Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.884250 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zs46w"] Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.886407 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.888977 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.889237 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.904596 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zs46w"] Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.915835 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:12 crc kubenswrapper[4799]: W1124 08:10:12.918343 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c6d5ad6_52a2_4a9e_a86a_2150e9d22667.slice/crio-8ff09599d8f354db3edf8890f99702c5c1064fa48e114253f0f9d29c354b37b8 WatchSource:0}: Error finding container 8ff09599d8f354db3edf8890f99702c5c1064fa48e114253f0f9d29c354b37b8: Status 404 returned error can't find the container with id 8ff09599d8f354db3edf8890f99702c5c1064fa48e114253f0f9d29c354b37b8 Nov 24 08:10:12 crc kubenswrapper[4799]: W1124 08:10:12.926129 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e9c42d8_a576_460f_8629_2536bd6efeba.slice/crio-bedea8514d232c85fb8dafa8b9fb80fcb4e62042818182c68cf0144bf4fbadce WatchSource:0}: Error finding container bedea8514d232c85fb8dafa8b9fb80fcb4e62042818182c68cf0144bf4fbadce: Status 404 returned error can't find the container with id bedea8514d232c85fb8dafa8b9fb80fcb4e62042818182c68cf0144bf4fbadce Nov 24 08:10:12 crc kubenswrapper[4799]: I1124 08:10:12.929377 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.048743 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.078227 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.081177 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q59s2\" (UniqueName: \"kubernetes.io/projected/97bf22b0-39f1-4302-947a-3e5ebbb078f5-kube-api-access-q59s2\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.081289 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-scripts\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.081330 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-config-data\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.152610 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:13 crc kubenswrapper[4799]: W1124 08:10:13.164592 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67e6baf3_e623_4860_9c7f_46cdd5d0d951.slice/crio-5a054e55ce086bfa300f54b9c8cac523ded8269d304c79560cc94559ae5f3534 WatchSource:0}: Error finding container 5a054e55ce086bfa300f54b9c8cac523ded8269d304c79560cc94559ae5f3534: Status 404 returned error can't find the container with id 5a054e55ce086bfa300f54b9c8cac523ded8269d304c79560cc94559ae5f3534 Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.175612 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cb57d6b9-bmvcf"] Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.184257 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.184430 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q59s2\" (UniqueName: \"kubernetes.io/projected/97bf22b0-39f1-4302-947a-3e5ebbb078f5-kube-api-access-q59s2\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.184557 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-scripts\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.184624 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-config-data\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.195964 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-config-data\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.199325 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-scripts\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.206042 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.217877 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q59s2\" (UniqueName: \"kubernetes.io/projected/97bf22b0-39f1-4302-947a-3e5ebbb078f5-kube-api-access-q59s2\") pod \"nova-cell1-conductor-db-sync-zs46w\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.337956 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.547281 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e9c42d8-a576-460f-8629-2536bd6efeba","Type":"ContainerStarted","Data":"bedea8514d232c85fb8dafa8b9fb80fcb4e62042818182c68cf0144bf4fbadce"} Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.551878 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" event={"ID":"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090","Type":"ContainerStarted","Data":"71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf"} Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.551942 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" event={"ID":"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090","Type":"ContainerStarted","Data":"15b3c6b09e17acb9871512d866c0cbe7ea94c310b2e315f0429739f2312330a9"} Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.555989 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667","Type":"ContainerStarted","Data":"8ff09599d8f354db3edf8890f99702c5c1064fa48e114253f0f9d29c354b37b8"} Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.562130 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-72q6b" event={"ID":"d1dc5c43-814a-4a95-91b2-75880697ab67","Type":"ContainerStarted","Data":"7152c16ef8ee65b421528af336df1b62c92830cbb7762ae412eaa4fe8d89eb4b"} Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.562185 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-72q6b" event={"ID":"d1dc5c43-814a-4a95-91b2-75880697ab67","Type":"ContainerStarted","Data":"effcb61e1fcfbaa811204450fa35f8bbda2c328180da9a4c2997f342df7a8505"} Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.566577 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"67e6baf3-e623-4860-9c7f-46cdd5d0d951","Type":"ContainerStarted","Data":"5a054e55ce086bfa300f54b9c8cac523ded8269d304c79560cc94559ae5f3534"} Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.582715 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"35448c34-84e0-4cbf-aec7-a9c3d3adb04e","Type":"ContainerStarted","Data":"779b088622649e15cc88b178d91a5a83e8def85967270ceb014e65a345325838"} Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.623729 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-72q6b" podStartSLOduration=2.623478031 podStartE2EDuration="2.623478031s" podCreationTimestamp="2025-11-24 08:10:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:13.59320284 +0000 UTC m=+4959.249185314" watchObservedRunningTime="2025-11-24 08:10:13.623478031 +0000 UTC m=+4959.279460505" Nov 24 08:10:13 crc kubenswrapper[4799]: I1124 08:10:13.939918 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zs46w"] Nov 24 08:10:14 crc kubenswrapper[4799]: I1124 08:10:14.599414 4799 generic.go:334] "Generic (PLEG): container finished" podID="1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" containerID="71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf" exitCode=0 Nov 24 08:10:14 crc kubenswrapper[4799]: I1124 08:10:14.600898 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" event={"ID":"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090","Type":"ContainerDied","Data":"71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf"} Nov 24 08:10:14 crc kubenswrapper[4799]: I1124 08:10:14.600945 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" event={"ID":"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090","Type":"ContainerStarted","Data":"10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14"} Nov 24 08:10:14 crc kubenswrapper[4799]: I1124 08:10:14.600964 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:14 crc kubenswrapper[4799]: I1124 08:10:14.627788 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" podStartSLOduration=2.627770924 podStartE2EDuration="2.627770924s" podCreationTimestamp="2025-11-24 08:10:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:14.618712633 +0000 UTC m=+4960.274695117" watchObservedRunningTime="2025-11-24 08:10:14.627770924 +0000 UTC m=+4960.283753398" Nov 24 08:10:15 crc kubenswrapper[4799]: I1124 08:10:15.623539 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zs46w" event={"ID":"97bf22b0-39f1-4302-947a-3e5ebbb078f5","Type":"ContainerStarted","Data":"834874093e3077920809f48c502086e476f9c6292cb527dbdc0cb04d63fb0057"} Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.640472 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"67e6baf3-e623-4860-9c7f-46cdd5d0d951","Type":"ContainerStarted","Data":"49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d"} Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.640858 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"67e6baf3-e623-4860-9c7f-46cdd5d0d951","Type":"ContainerStarted","Data":"b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0"} Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.646815 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"35448c34-84e0-4cbf-aec7-a9c3d3adb04e","Type":"ContainerStarted","Data":"d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb"} Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.649538 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zs46w" event={"ID":"97bf22b0-39f1-4302-947a-3e5ebbb078f5","Type":"ContainerStarted","Data":"c03b6b7128b540198200d3ac04c9b57ab126615ec725dd3c258478c1e31a49ec"} Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.654323 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e9c42d8-a576-460f-8629-2536bd6efeba","Type":"ContainerStarted","Data":"6f1ec4e6c4d0d6987d0877ab3b001517e6bd2387e06ef3e942cffe61e5fb34bc"} Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.657884 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667","Type":"ContainerStarted","Data":"ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37"} Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.657920 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667","Type":"ContainerStarted","Data":"fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666"} Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.673000 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.29464615 podStartE2EDuration="5.67297582s" podCreationTimestamp="2025-11-24 08:10:11 +0000 UTC" firstStartedPulling="2025-11-24 08:10:13.169685753 +0000 UTC m=+4958.825668227" lastFinishedPulling="2025-11-24 08:10:15.548015423 +0000 UTC m=+4961.203997897" observedRunningTime="2025-11-24 08:10:16.668384552 +0000 UTC m=+4962.324367036" watchObservedRunningTime="2025-11-24 08:10:16.67297582 +0000 UTC m=+4962.328958304" Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.691257 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-zs46w" podStartSLOduration=4.690823496 podStartE2EDuration="4.690823496s" podCreationTimestamp="2025-11-24 08:10:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:16.686745322 +0000 UTC m=+4962.342727796" watchObservedRunningTime="2025-11-24 08:10:16.690823496 +0000 UTC m=+4962.346805990" Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.754192 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.139118819 podStartE2EDuration="5.75396203s" podCreationTimestamp="2025-11-24 08:10:11 +0000 UTC" firstStartedPulling="2025-11-24 08:10:12.929980043 +0000 UTC m=+4958.585962507" lastFinishedPulling="2025-11-24 08:10:15.544823254 +0000 UTC m=+4961.200805718" observedRunningTime="2025-11-24 08:10:16.74928269 +0000 UTC m=+4962.405265174" watchObservedRunningTime="2025-11-24 08:10:16.75396203 +0000 UTC m=+4962.409944504" Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.763949 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.14240518 podStartE2EDuration="5.763928477s" podCreationTimestamp="2025-11-24 08:10:11 +0000 UTC" firstStartedPulling="2025-11-24 08:10:12.921627131 +0000 UTC m=+4958.577609605" lastFinishedPulling="2025-11-24 08:10:15.543150428 +0000 UTC m=+4961.199132902" observedRunningTime="2025-11-24 08:10:16.731490436 +0000 UTC m=+4962.387472910" watchObservedRunningTime="2025-11-24 08:10:16.763928477 +0000 UTC m=+4962.419910951" Nov 24 08:10:16 crc kubenswrapper[4799]: I1124 08:10:16.780707 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.279765897 podStartE2EDuration="5.780684613s" podCreationTimestamp="2025-11-24 08:10:11 +0000 UTC" firstStartedPulling="2025-11-24 08:10:13.043360163 +0000 UTC m=+4958.699342637" lastFinishedPulling="2025-11-24 08:10:15.544278879 +0000 UTC m=+4961.200261353" observedRunningTime="2025-11-24 08:10:16.773527474 +0000 UTC m=+4962.429509958" watchObservedRunningTime="2025-11-24 08:10:16.780684613 +0000 UTC m=+4962.436667087" Nov 24 08:10:17 crc kubenswrapper[4799]: I1124 08:10:17.260182 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:17 crc kubenswrapper[4799]: I1124 08:10:17.279023 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:10:17 crc kubenswrapper[4799]: I1124 08:10:17.279101 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:10:17 crc kubenswrapper[4799]: I1124 08:10:17.387271 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 08:10:18 crc kubenswrapper[4799]: I1124 08:10:18.687878 4799 generic.go:334] "Generic (PLEG): container finished" podID="97bf22b0-39f1-4302-947a-3e5ebbb078f5" containerID="c03b6b7128b540198200d3ac04c9b57ab126615ec725dd3c258478c1e31a49ec" exitCode=0 Nov 24 08:10:18 crc kubenswrapper[4799]: I1124 08:10:18.687983 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zs46w" event={"ID":"97bf22b0-39f1-4302-947a-3e5ebbb078f5","Type":"ContainerDied","Data":"c03b6b7128b540198200d3ac04c9b57ab126615ec725dd3c258478c1e31a49ec"} Nov 24 08:10:18 crc kubenswrapper[4799]: I1124 08:10:18.690639 4799 generic.go:334] "Generic (PLEG): container finished" podID="d1dc5c43-814a-4a95-91b2-75880697ab67" containerID="7152c16ef8ee65b421528af336df1b62c92830cbb7762ae412eaa4fe8d89eb4b" exitCode=0 Nov 24 08:10:18 crc kubenswrapper[4799]: I1124 08:10:18.690821 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-72q6b" event={"ID":"d1dc5c43-814a-4a95-91b2-75880697ab67","Type":"ContainerDied","Data":"7152c16ef8ee65b421528af336df1b62c92830cbb7762ae412eaa4fe8d89eb4b"} Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.082834 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.169418 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.222513 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-scripts\") pod \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.222564 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-combined-ca-bundle\") pod \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.222600 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q59s2\" (UniqueName: \"kubernetes.io/projected/97bf22b0-39f1-4302-947a-3e5ebbb078f5-kube-api-access-q59s2\") pod \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.222643 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-config-data\") pod \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\" (UID: \"97bf22b0-39f1-4302-947a-3e5ebbb078f5\") " Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.228500 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-scripts" (OuterVolumeSpecName: "scripts") pod "97bf22b0-39f1-4302-947a-3e5ebbb078f5" (UID: "97bf22b0-39f1-4302-947a-3e5ebbb078f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.229347 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97bf22b0-39f1-4302-947a-3e5ebbb078f5-kube-api-access-q59s2" (OuterVolumeSpecName: "kube-api-access-q59s2") pod "97bf22b0-39f1-4302-947a-3e5ebbb078f5" (UID: "97bf22b0-39f1-4302-947a-3e5ebbb078f5"). InnerVolumeSpecName "kube-api-access-q59s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.250080 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-config-data" (OuterVolumeSpecName: "config-data") pod "97bf22b0-39f1-4302-947a-3e5ebbb078f5" (UID: "97bf22b0-39f1-4302-947a-3e5ebbb078f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.251488 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97bf22b0-39f1-4302-947a-3e5ebbb078f5" (UID: "97bf22b0-39f1-4302-947a-3e5ebbb078f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.325245 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2kmf\" (UniqueName: \"kubernetes.io/projected/d1dc5c43-814a-4a95-91b2-75880697ab67-kube-api-access-z2kmf\") pod \"d1dc5c43-814a-4a95-91b2-75880697ab67\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.325290 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-config-data\") pod \"d1dc5c43-814a-4a95-91b2-75880697ab67\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.325396 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-combined-ca-bundle\") pod \"d1dc5c43-814a-4a95-91b2-75880697ab67\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.325504 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-scripts\") pod \"d1dc5c43-814a-4a95-91b2-75880697ab67\" (UID: \"d1dc5c43-814a-4a95-91b2-75880697ab67\") " Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.325870 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.325900 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.325916 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q59s2\" (UniqueName: \"kubernetes.io/projected/97bf22b0-39f1-4302-947a-3e5ebbb078f5-kube-api-access-q59s2\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.325932 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97bf22b0-39f1-4302-947a-3e5ebbb078f5-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.328478 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1dc5c43-814a-4a95-91b2-75880697ab67-kube-api-access-z2kmf" (OuterVolumeSpecName: "kube-api-access-z2kmf") pod "d1dc5c43-814a-4a95-91b2-75880697ab67" (UID: "d1dc5c43-814a-4a95-91b2-75880697ab67"). InnerVolumeSpecName "kube-api-access-z2kmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.328746 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-scripts" (OuterVolumeSpecName: "scripts") pod "d1dc5c43-814a-4a95-91b2-75880697ab67" (UID: "d1dc5c43-814a-4a95-91b2-75880697ab67"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.348150 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1dc5c43-814a-4a95-91b2-75880697ab67" (UID: "d1dc5c43-814a-4a95-91b2-75880697ab67"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.348547 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-config-data" (OuterVolumeSpecName: "config-data") pod "d1dc5c43-814a-4a95-91b2-75880697ab67" (UID: "d1dc5c43-814a-4a95-91b2-75880697ab67"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.400725 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.400792 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.400832 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.401349 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.401410 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" gracePeriod=600 Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.428056 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.428089 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.428107 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1dc5c43-814a-4a95-91b2-75880697ab67-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.428126 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2kmf\" (UniqueName: \"kubernetes.io/projected/d1dc5c43-814a-4a95-91b2-75880697ab67-kube-api-access-z2kmf\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:20 crc kubenswrapper[4799]: E1124 08:10:20.522491 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.710763 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-72q6b" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.710773 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-72q6b" event={"ID":"d1dc5c43-814a-4a95-91b2-75880697ab67","Type":"ContainerDied","Data":"effcb61e1fcfbaa811204450fa35f8bbda2c328180da9a4c2997f342df7a8505"} Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.711542 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="effcb61e1fcfbaa811204450fa35f8bbda2c328180da9a4c2997f342df7a8505" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.717044 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" exitCode=0 Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.717111 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329"} Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.717143 4799 scope.go:117] "RemoveContainer" containerID="023b79a79ebb461005d32f47310810404d347caf336f616e7c3a30dfcc69e291" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.717732 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:10:20 crc kubenswrapper[4799]: E1124 08:10:20.718083 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.719480 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zs46w" event={"ID":"97bf22b0-39f1-4302-947a-3e5ebbb078f5","Type":"ContainerDied","Data":"834874093e3077920809f48c502086e476f9c6292cb527dbdc0cb04d63fb0057"} Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.719511 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="834874093e3077920809f48c502086e476f9c6292cb527dbdc0cb04d63fb0057" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.719558 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zs46w" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.856974 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:10:20 crc kubenswrapper[4799]: E1124 08:10:20.857356 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1dc5c43-814a-4a95-91b2-75880697ab67" containerName="nova-manage" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.857382 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1dc5c43-814a-4a95-91b2-75880697ab67" containerName="nova-manage" Nov 24 08:10:20 crc kubenswrapper[4799]: E1124 08:10:20.857408 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97bf22b0-39f1-4302-947a-3e5ebbb078f5" containerName="nova-cell1-conductor-db-sync" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.857415 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="97bf22b0-39f1-4302-947a-3e5ebbb078f5" containerName="nova-cell1-conductor-db-sync" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.857583 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="97bf22b0-39f1-4302-947a-3e5ebbb078f5" containerName="nova-cell1-conductor-db-sync" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.857605 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1dc5c43-814a-4a95-91b2-75880697ab67" containerName="nova-manage" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.858287 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.861322 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.866785 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.937617 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.938069 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.938281 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdkq5\" (UniqueName: \"kubernetes.io/projected/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-kube-api-access-hdkq5\") pod \"nova-cell1-conductor-0\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.958363 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.958904 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="67e6baf3-e623-4860-9c7f-46cdd5d0d951" containerName="nova-api-log" containerID="cri-o://b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0" gracePeriod=30 Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.958963 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="67e6baf3-e623-4860-9c7f-46cdd5d0d951" containerName="nova-api-api" containerID="cri-o://49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d" gracePeriod=30 Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.967244 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.967441 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="35448c34-84e0-4cbf-aec7-a9c3d3adb04e" containerName="nova-scheduler-scheduler" containerID="cri-o://d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb" gracePeriod=30 Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.983647 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.983924 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" containerName="nova-metadata-log" containerID="cri-o://fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666" gracePeriod=30 Nov 24 08:10:20 crc kubenswrapper[4799]: I1124 08:10:20.984121 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" containerName="nova-metadata-metadata" containerID="cri-o://ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37" gracePeriod=30 Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.040466 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.040529 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdkq5\" (UniqueName: \"kubernetes.io/projected/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-kube-api-access-hdkq5\") pod \"nova-cell1-conductor-0\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.040588 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.044400 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.044918 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.059742 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdkq5\" (UniqueName: \"kubernetes.io/projected/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-kube-api-access-hdkq5\") pod \"nova-cell1-conductor-0\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.182198 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.510652 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.570407 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.655474 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-combined-ca-bundle\") pod \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.655542 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-config-data\") pod \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.655648 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc4cf\" (UniqueName: \"kubernetes.io/projected/67e6baf3-e623-4860-9c7f-46cdd5d0d951-kube-api-access-wc4cf\") pod \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.655706 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-logs\") pod \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.655727 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67e6baf3-e623-4860-9c7f-46cdd5d0d951-logs\") pod \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.655769 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-config-data\") pod \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.655788 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-combined-ca-bundle\") pod \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\" (UID: \"67e6baf3-e623-4860-9c7f-46cdd5d0d951\") " Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.655877 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wp6p\" (UniqueName: \"kubernetes.io/projected/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-kube-api-access-2wp6p\") pod \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\" (UID: \"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667\") " Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.656039 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67e6baf3-e623-4860-9c7f-46cdd5d0d951-logs" (OuterVolumeSpecName: "logs") pod "67e6baf3-e623-4860-9c7f-46cdd5d0d951" (UID: "67e6baf3-e623-4860-9c7f-46cdd5d0d951"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.656237 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-logs" (OuterVolumeSpecName: "logs") pod "7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" (UID: "7c6d5ad6-52a2-4a9e-a86a-2150e9d22667"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.656290 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67e6baf3-e623-4860-9c7f-46cdd5d0d951-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.661146 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e6baf3-e623-4860-9c7f-46cdd5d0d951-kube-api-access-wc4cf" (OuterVolumeSpecName: "kube-api-access-wc4cf") pod "67e6baf3-e623-4860-9c7f-46cdd5d0d951" (UID: "67e6baf3-e623-4860-9c7f-46cdd5d0d951"). InnerVolumeSpecName "kube-api-access-wc4cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.661312 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-kube-api-access-2wp6p" (OuterVolumeSpecName: "kube-api-access-2wp6p") pod "7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" (UID: "7c6d5ad6-52a2-4a9e-a86a-2150e9d22667"). InnerVolumeSpecName "kube-api-access-2wp6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.685562 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-config-data" (OuterVolumeSpecName: "config-data") pod "7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" (UID: "7c6d5ad6-52a2-4a9e-a86a-2150e9d22667"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.687451 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-config-data" (OuterVolumeSpecName: "config-data") pod "67e6baf3-e623-4860-9c7f-46cdd5d0d951" (UID: "67e6baf3-e623-4860-9c7f-46cdd5d0d951"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.689616 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67e6baf3-e623-4860-9c7f-46cdd5d0d951" (UID: "67e6baf3-e623-4860-9c7f-46cdd5d0d951"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.693270 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" (UID: "7c6d5ad6-52a2-4a9e-a86a-2150e9d22667"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:21 crc kubenswrapper[4799]: W1124 08:10:21.716241 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d01fe98_c3d6_45b2_83a8_4ef1f00ec3e7.slice/crio-7899e49b55df1c5e848aab827269fac06fb78ab49880d1621fb4b6dfc1c3cbce WatchSource:0}: Error finding container 7899e49b55df1c5e848aab827269fac06fb78ab49880d1621fb4b6dfc1c3cbce: Status 404 returned error can't find the container with id 7899e49b55df1c5e848aab827269fac06fb78ab49880d1621fb4b6dfc1c3cbce Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.717180 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.733816 4799 generic.go:334] "Generic (PLEG): container finished" podID="67e6baf3-e623-4860-9c7f-46cdd5d0d951" containerID="49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d" exitCode=0 Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.733870 4799 generic.go:334] "Generic (PLEG): container finished" podID="67e6baf3-e623-4860-9c7f-46cdd5d0d951" containerID="b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0" exitCode=143 Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.733895 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.733864 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"67e6baf3-e623-4860-9c7f-46cdd5d0d951","Type":"ContainerDied","Data":"49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d"} Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.733946 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"67e6baf3-e623-4860-9c7f-46cdd5d0d951","Type":"ContainerDied","Data":"b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0"} Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.733960 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"67e6baf3-e623-4860-9c7f-46cdd5d0d951","Type":"ContainerDied","Data":"5a054e55ce086bfa300f54b9c8cac523ded8269d304c79560cc94559ae5f3534"} Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.733981 4799 scope.go:117] "RemoveContainer" containerID="49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.736098 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7","Type":"ContainerStarted","Data":"7899e49b55df1c5e848aab827269fac06fb78ab49880d1621fb4b6dfc1c3cbce"} Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.739023 4799 generic.go:334] "Generic (PLEG): container finished" podID="7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" containerID="ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37" exitCode=0 Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.739053 4799 generic.go:334] "Generic (PLEG): container finished" podID="7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" containerID="fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666" exitCode=143 Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.739076 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667","Type":"ContainerDied","Data":"ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37"} Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.739109 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667","Type":"ContainerDied","Data":"fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666"} Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.739122 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7c6d5ad6-52a2-4a9e-a86a-2150e9d22667","Type":"ContainerDied","Data":"8ff09599d8f354db3edf8890f99702c5c1064fa48e114253f0f9d29c354b37b8"} Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.739193 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.757512 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc4cf\" (UniqueName: \"kubernetes.io/projected/67e6baf3-e623-4860-9c7f-46cdd5d0d951-kube-api-access-wc4cf\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.757533 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.757542 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.757550 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e6baf3-e623-4860-9c7f-46cdd5d0d951-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.757560 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wp6p\" (UniqueName: \"kubernetes.io/projected/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-kube-api-access-2wp6p\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.757568 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.757577 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.769197 4799 scope.go:117] "RemoveContainer" containerID="b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.800592 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.815249 4799 scope.go:117] "RemoveContainer" containerID="49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d" Nov 24 08:10:21 crc kubenswrapper[4799]: E1124 08:10:21.828290 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d\": container with ID starting with 49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d not found: ID does not exist" containerID="49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.828339 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d"} err="failed to get container status \"49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d\": rpc error: code = NotFound desc = could not find container \"49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d\": container with ID starting with 49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d not found: ID does not exist" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.828368 4799 scope.go:117] "RemoveContainer" containerID="b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.830632 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:21 crc kubenswrapper[4799]: E1124 08:10:21.831480 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0\": container with ID starting with b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0 not found: ID does not exist" containerID="b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.831633 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0"} err="failed to get container status \"b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0\": rpc error: code = NotFound desc = could not find container \"b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0\": container with ID starting with b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0 not found: ID does not exist" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.831704 4799 scope.go:117] "RemoveContainer" containerID="49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.832468 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d"} err="failed to get container status \"49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d\": rpc error: code = NotFound desc = could not find container \"49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d\": container with ID starting with 49388664cfbf82a220ded85958b830b2bdc094c0238ca9dc29687ec89623790d not found: ID does not exist" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.832508 4799 scope.go:117] "RemoveContainer" containerID="b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.834658 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0"} err="failed to get container status \"b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0\": rpc error: code = NotFound desc = could not find container \"b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0\": container with ID starting with b5e1d326ae161958a39d997a519901ba252e94c059ff02fec1c8a259925beea0 not found: ID does not exist" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.834703 4799 scope.go:117] "RemoveContainer" containerID="ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.842029 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.867429 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.875838 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:21 crc kubenswrapper[4799]: E1124 08:10:21.876309 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e6baf3-e623-4860-9c7f-46cdd5d0d951" containerName="nova-api-api" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.876332 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e6baf3-e623-4860-9c7f-46cdd5d0d951" containerName="nova-api-api" Nov 24 08:10:21 crc kubenswrapper[4799]: E1124 08:10:21.876349 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" containerName="nova-metadata-metadata" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.876359 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" containerName="nova-metadata-metadata" Nov 24 08:10:21 crc kubenswrapper[4799]: E1124 08:10:21.876375 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e6baf3-e623-4860-9c7f-46cdd5d0d951" containerName="nova-api-log" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.876384 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e6baf3-e623-4860-9c7f-46cdd5d0d951" containerName="nova-api-log" Nov 24 08:10:21 crc kubenswrapper[4799]: E1124 08:10:21.876408 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" containerName="nova-metadata-log" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.876417 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" containerName="nova-metadata-log" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.876569 4799 scope.go:117] "RemoveContainer" containerID="fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.876650 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" containerName="nova-metadata-metadata" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.876674 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e6baf3-e623-4860-9c7f-46cdd5d0d951" containerName="nova-api-log" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.876694 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e6baf3-e623-4860-9c7f-46cdd5d0d951" containerName="nova-api-api" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.876721 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" containerName="nova-metadata-log" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.877867 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.880699 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.885002 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:21 crc kubenswrapper[4799]: E1124 08:10:21.887514 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67e6baf3_e623_4860_9c7f_46cdd5d0d951.slice\": RecentStats: unable to find data in memory cache]" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.900121 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.902807 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.905026 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.908956 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.919283 4799 scope.go:117] "RemoveContainer" containerID="ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37" Nov 24 08:10:21 crc kubenswrapper[4799]: E1124 08:10:21.919792 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37\": container with ID starting with ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37 not found: ID does not exist" containerID="ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.919859 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37"} err="failed to get container status \"ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37\": rpc error: code = NotFound desc = could not find container \"ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37\": container with ID starting with ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37 not found: ID does not exist" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.919893 4799 scope.go:117] "RemoveContainer" containerID="fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666" Nov 24 08:10:21 crc kubenswrapper[4799]: E1124 08:10:21.920290 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666\": container with ID starting with fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666 not found: ID does not exist" containerID="fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.920323 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666"} err="failed to get container status \"fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666\": rpc error: code = NotFound desc = could not find container \"fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666\": container with ID starting with fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666 not found: ID does not exist" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.920350 4799 scope.go:117] "RemoveContainer" containerID="ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.920808 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37"} err="failed to get container status \"ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37\": rpc error: code = NotFound desc = could not find container \"ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37\": container with ID starting with ef74374038cb14cbcb4f65e207de80a22b30dca49cb9ba2a949e33b20fe43d37 not found: ID does not exist" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.920828 4799 scope.go:117] "RemoveContainer" containerID="fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.922170 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666"} err="failed to get container status \"fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666\": rpc error: code = NotFound desc = could not find container \"fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666\": container with ID starting with fe8dbf68fc253eb118b23d2d05417ac866e2dde044f13e54dd79bb320b543666 not found: ID does not exist" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.967466 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-config-data\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.967711 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvbhs\" (UniqueName: \"kubernetes.io/projected/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-kube-api-access-zvbhs\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.967945 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-logs\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:21 crc kubenswrapper[4799]: I1124 08:10:21.968387 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.069646 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.069696 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-config-data\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.069713 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvbhs\" (UniqueName: \"kubernetes.io/projected/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-kube-api-access-zvbhs\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.069741 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42256f8c-898d-43b7-ac4c-bb98095b5702-logs\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.069807 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.069903 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-logs\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.069930 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb5hw\" (UniqueName: \"kubernetes.io/projected/42256f8c-898d-43b7-ac4c-bb98095b5702-kube-api-access-hb5hw\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.069949 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-config-data\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.070356 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-logs\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.074200 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-config-data\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.074291 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.086790 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvbhs\" (UniqueName: \"kubernetes.io/projected/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-kube-api-access-zvbhs\") pod \"nova-metadata-0\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " pod="openstack/nova-metadata-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.171482 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb5hw\" (UniqueName: \"kubernetes.io/projected/42256f8c-898d-43b7-ac4c-bb98095b5702-kube-api-access-hb5hw\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.171565 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-config-data\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.171707 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42256f8c-898d-43b7-ac4c-bb98095b5702-logs\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.171833 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.173812 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42256f8c-898d-43b7-ac4c-bb98095b5702-logs\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.175764 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-config-data\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.175894 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.199791 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb5hw\" (UniqueName: \"kubernetes.io/projected/42256f8c-898d-43b7-ac4c-bb98095b5702-kube-api-access-hb5hw\") pod \"nova-api-0\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.203433 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.225629 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.260253 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.280481 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.523010 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.583022 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5dc9fb77-h8l9n"] Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.583469 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" podUID="3f8cae7e-b65e-4104-86b8-f2c6fa283f35" containerName="dnsmasq-dns" containerID="cri-o://d16ac7e0747a57083af24878e129a3b8295cdef68f6df2dea9e2c771a62c8095" gracePeriod=10 Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.695161 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:22 crc kubenswrapper[4799]: W1124 08:10:22.698619 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42256f8c_898d_43b7_ac4c_bb98095b5702.slice/crio-72badaebe45f6f77e4bdd4057af93b9a1429a5eede3ca3e59fcd39fb54839d61 WatchSource:0}: Error finding container 72badaebe45f6f77e4bdd4057af93b9a1429a5eede3ca3e59fcd39fb54839d61: Status 404 returned error can't find the container with id 72badaebe45f6f77e4bdd4057af93b9a1429a5eede3ca3e59fcd39fb54839d61 Nov 24 08:10:22 crc kubenswrapper[4799]: W1124 08:10:22.702632 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cd4d364_cdfe_4a69_bf54_9858afa7b99e.slice/crio-1591b633b17bdcf9ee88610d35bca020850bbe14fd1113db3e9e1ee609a3708e WatchSource:0}: Error finding container 1591b633b17bdcf9ee88610d35bca020850bbe14fd1113db3e9e1ee609a3708e: Status 404 returned error can't find the container with id 1591b633b17bdcf9ee88610d35bca020850bbe14fd1113db3e9e1ee609a3708e Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.710441 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.753034 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42256f8c-898d-43b7-ac4c-bb98095b5702","Type":"ContainerStarted","Data":"72badaebe45f6f77e4bdd4057af93b9a1429a5eede3ca3e59fcd39fb54839d61"} Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.756675 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cd4d364-cdfe-4a69-bf54-9858afa7b99e","Type":"ContainerStarted","Data":"1591b633b17bdcf9ee88610d35bca020850bbe14fd1113db3e9e1ee609a3708e"} Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.758598 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7","Type":"ContainerStarted","Data":"8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511"} Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.758960 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.770469 4799 generic.go:334] "Generic (PLEG): container finished" podID="3f8cae7e-b65e-4104-86b8-f2c6fa283f35" containerID="d16ac7e0747a57083af24878e129a3b8295cdef68f6df2dea9e2c771a62c8095" exitCode=0 Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.770525 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" event={"ID":"3f8cae7e-b65e-4104-86b8-f2c6fa283f35","Type":"ContainerDied","Data":"d16ac7e0747a57083af24878e129a3b8295cdef68f6df2dea9e2c771a62c8095"} Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.781737 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.781718968 podStartE2EDuration="2.781718968s" podCreationTimestamp="2025-11-24 08:10:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:22.774821896 +0000 UTC m=+4968.430804370" watchObservedRunningTime="2025-11-24 08:10:22.781718968 +0000 UTC m=+4968.437701442" Nov 24 08:10:22 crc kubenswrapper[4799]: I1124 08:10:22.785539 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.283945 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.399441 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-config\") pod \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.399573 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-dns-svc\") pod \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.399630 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-sb\") pod \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.399670 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85pb4\" (UniqueName: \"kubernetes.io/projected/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-kube-api-access-85pb4\") pod \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.399695 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-nb\") pod \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\" (UID: \"3f8cae7e-b65e-4104-86b8-f2c6fa283f35\") " Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.415090 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-kube-api-access-85pb4" (OuterVolumeSpecName: "kube-api-access-85pb4") pod "3f8cae7e-b65e-4104-86b8-f2c6fa283f35" (UID: "3f8cae7e-b65e-4104-86b8-f2c6fa283f35"). InnerVolumeSpecName "kube-api-access-85pb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.479292 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3f8cae7e-b65e-4104-86b8-f2c6fa283f35" (UID: "3f8cae7e-b65e-4104-86b8-f2c6fa283f35"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.491993 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3f8cae7e-b65e-4104-86b8-f2c6fa283f35" (UID: "3f8cae7e-b65e-4104-86b8-f2c6fa283f35"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.493758 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3f8cae7e-b65e-4104-86b8-f2c6fa283f35" (UID: "3f8cae7e-b65e-4104-86b8-f2c6fa283f35"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.501810 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.501855 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.501871 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85pb4\" (UniqueName: \"kubernetes.io/projected/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-kube-api-access-85pb4\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.501885 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.505969 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-config" (OuterVolumeSpecName: "config") pod "3f8cae7e-b65e-4104-86b8-f2c6fa283f35" (UID: "3f8cae7e-b65e-4104-86b8-f2c6fa283f35"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.534330 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.604120 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f8cae7e-b65e-4104-86b8-f2c6fa283f35-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.639296 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67e6baf3-e623-4860-9c7f-46cdd5d0d951" path="/var/lib/kubelet/pods/67e6baf3-e623-4860-9c7f-46cdd5d0d951/volumes" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.640033 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c6d5ad6-52a2-4a9e-a86a-2150e9d22667" path="/var/lib/kubelet/pods/7c6d5ad6-52a2-4a9e-a86a-2150e9d22667/volumes" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.706013 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-config-data\") pod \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.706103 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bpck\" (UniqueName: \"kubernetes.io/projected/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-kube-api-access-4bpck\") pod \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.706773 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-combined-ca-bundle\") pod \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\" (UID: \"35448c34-84e0-4cbf-aec7-a9c3d3adb04e\") " Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.709421 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-kube-api-access-4bpck" (OuterVolumeSpecName: "kube-api-access-4bpck") pod "35448c34-84e0-4cbf-aec7-a9c3d3adb04e" (UID: "35448c34-84e0-4cbf-aec7-a9c3d3adb04e"). InnerVolumeSpecName "kube-api-access-4bpck". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.727518 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-config-data" (OuterVolumeSpecName: "config-data") pod "35448c34-84e0-4cbf-aec7-a9c3d3adb04e" (UID: "35448c34-84e0-4cbf-aec7-a9c3d3adb04e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.729059 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35448c34-84e0-4cbf-aec7-a9c3d3adb04e" (UID: "35448c34-84e0-4cbf-aec7-a9c3d3adb04e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.782127 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42256f8c-898d-43b7-ac4c-bb98095b5702","Type":"ContainerStarted","Data":"c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a"} Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.782173 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42256f8c-898d-43b7-ac4c-bb98095b5702","Type":"ContainerStarted","Data":"e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b"} Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.784757 4799 generic.go:334] "Generic (PLEG): container finished" podID="35448c34-84e0-4cbf-aec7-a9c3d3adb04e" containerID="d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb" exitCode=0 Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.784807 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"35448c34-84e0-4cbf-aec7-a9c3d3adb04e","Type":"ContainerDied","Data":"d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb"} Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.784869 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"35448c34-84e0-4cbf-aec7-a9c3d3adb04e","Type":"ContainerDied","Data":"779b088622649e15cc88b178d91a5a83e8def85967270ceb014e65a345325838"} Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.784892 4799 scope.go:117] "RemoveContainer" containerID="d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.785007 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.790560 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cd4d364-cdfe-4a69-bf54-9858afa7b99e","Type":"ContainerStarted","Data":"56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e"} Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.790606 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cd4d364-cdfe-4a69-bf54-9858afa7b99e","Type":"ContainerStarted","Data":"4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296"} Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.794074 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.794509 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5dc9fb77-h8l9n" event={"ID":"3f8cae7e-b65e-4104-86b8-f2c6fa283f35","Type":"ContainerDied","Data":"8ba29457ddfa81ff4b8f69ee286bd5612b88ba4b2b1a632bf3575978aa45483e"} Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.811430 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.811461 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.811471 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bpck\" (UniqueName: \"kubernetes.io/projected/35448c34-84e0-4cbf-aec7-a9c3d3adb04e-kube-api-access-4bpck\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.814338 4799 scope.go:117] "RemoveContainer" containerID="d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb" Nov 24 08:10:23 crc kubenswrapper[4799]: E1124 08:10:23.814813 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb\": container with ID starting with d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb not found: ID does not exist" containerID="d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.814893 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb"} err="failed to get container status \"d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb\": rpc error: code = NotFound desc = could not find container \"d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb\": container with ID starting with d32a2c476836504fd0d8b364d028d65157d4bed69780e72287b7ecb5f5680bdb not found: ID does not exist" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.814925 4799 scope.go:117] "RemoveContainer" containerID="d16ac7e0747a57083af24878e129a3b8295cdef68f6df2dea9e2c771a62c8095" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.817120 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.817100106 podStartE2EDuration="2.817100106s" podCreationTimestamp="2025-11-24 08:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:23.807469988 +0000 UTC m=+4969.463452472" watchObservedRunningTime="2025-11-24 08:10:23.817100106 +0000 UTC m=+4969.473082580" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.838661 4799 scope.go:117] "RemoveContainer" containerID="ef6a9d756bb3f1f53d16a4a16d620f93f94c1d614e5c9efa62b3e82697dc798f" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.841744 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.84172362 podStartE2EDuration="2.84172362s" podCreationTimestamp="2025-11-24 08:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:23.822800104 +0000 UTC m=+4969.478782598" watchObservedRunningTime="2025-11-24 08:10:23.84172362 +0000 UTC m=+4969.497706094" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.877319 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.889034 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.897098 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5dc9fb77-h8l9n"] Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.906140 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:23 crc kubenswrapper[4799]: E1124 08:10:23.906533 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35448c34-84e0-4cbf-aec7-a9c3d3adb04e" containerName="nova-scheduler-scheduler" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.906546 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="35448c34-84e0-4cbf-aec7-a9c3d3adb04e" containerName="nova-scheduler-scheduler" Nov 24 08:10:23 crc kubenswrapper[4799]: E1124 08:10:23.906577 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f8cae7e-b65e-4104-86b8-f2c6fa283f35" containerName="init" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.906584 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f8cae7e-b65e-4104-86b8-f2c6fa283f35" containerName="init" Nov 24 08:10:23 crc kubenswrapper[4799]: E1124 08:10:23.906591 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f8cae7e-b65e-4104-86b8-f2c6fa283f35" containerName="dnsmasq-dns" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.906597 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f8cae7e-b65e-4104-86b8-f2c6fa283f35" containerName="dnsmasq-dns" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.906753 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f8cae7e-b65e-4104-86b8-f2c6fa283f35" containerName="dnsmasq-dns" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.906766 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="35448c34-84e0-4cbf-aec7-a9c3d3adb04e" containerName="nova-scheduler-scheduler" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.907363 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.908919 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.917984 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c5dc9fb77-h8l9n"] Nov 24 08:10:23 crc kubenswrapper[4799]: I1124 08:10:23.928489 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.015138 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8xg2\" (UniqueName: \"kubernetes.io/projected/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-kube-api-access-q8xg2\") pod \"nova-scheduler-0\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.015191 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.015271 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-config-data\") pod \"nova-scheduler-0\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.116636 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8xg2\" (UniqueName: \"kubernetes.io/projected/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-kube-api-access-q8xg2\") pod \"nova-scheduler-0\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.116691 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.116747 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-config-data\") pod \"nova-scheduler-0\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.125729 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-config-data\") pod \"nova-scheduler-0\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.126211 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.144987 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8xg2\" (UniqueName: \"kubernetes.io/projected/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-kube-api-access-q8xg2\") pod \"nova-scheduler-0\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.225676 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.671592 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:24 crc kubenswrapper[4799]: W1124 08:10:24.680797 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ea30b20_ee93_452f_bf96_9bf7fbcddbe1.slice/crio-62c9003aeeb9316be2a9e8a67febf7d10c75f7e800ebfbd98a4cda8e14c017d5 WatchSource:0}: Error finding container 62c9003aeeb9316be2a9e8a67febf7d10c75f7e800ebfbd98a4cda8e14c017d5: Status 404 returned error can't find the container with id 62c9003aeeb9316be2a9e8a67febf7d10c75f7e800ebfbd98a4cda8e14c017d5 Nov 24 08:10:24 crc kubenswrapper[4799]: I1124 08:10:24.802359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1","Type":"ContainerStarted","Data":"62c9003aeeb9316be2a9e8a67febf7d10c75f7e800ebfbd98a4cda8e14c017d5"} Nov 24 08:10:25 crc kubenswrapper[4799]: I1124 08:10:25.644224 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35448c34-84e0-4cbf-aec7-a9c3d3adb04e" path="/var/lib/kubelet/pods/35448c34-84e0-4cbf-aec7-a9c3d3adb04e/volumes" Nov 24 08:10:25 crc kubenswrapper[4799]: I1124 08:10:25.645253 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f8cae7e-b65e-4104-86b8-f2c6fa283f35" path="/var/lib/kubelet/pods/3f8cae7e-b65e-4104-86b8-f2c6fa283f35/volumes" Nov 24 08:10:25 crc kubenswrapper[4799]: I1124 08:10:25.815783 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1","Type":"ContainerStarted","Data":"4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885"} Nov 24 08:10:25 crc kubenswrapper[4799]: I1124 08:10:25.849728 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.84969928 podStartE2EDuration="2.84969928s" podCreationTimestamp="2025-11-24 08:10:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:25.834537809 +0000 UTC m=+4971.490520303" watchObservedRunningTime="2025-11-24 08:10:25.84969928 +0000 UTC m=+4971.505681754" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.229138 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.743821 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-b8cxs"] Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.745258 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.747100 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.748042 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.765077 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-b8cxs"] Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.881233 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.881366 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-scripts\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.881602 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6ssz\" (UniqueName: \"kubernetes.io/projected/30a2e74b-c94d-448a-9d98-a0cc2539be3a-kube-api-access-v6ssz\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.881805 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-config-data\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.983665 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-config-data\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.983725 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.983934 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-scripts\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.983995 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6ssz\" (UniqueName: \"kubernetes.io/projected/30a2e74b-c94d-448a-9d98-a0cc2539be3a-kube-api-access-v6ssz\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:26 crc kubenswrapper[4799]: I1124 08:10:26.995418 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:27 crc kubenswrapper[4799]: I1124 08:10:27.001719 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-config-data\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:27 crc kubenswrapper[4799]: I1124 08:10:27.008575 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6ssz\" (UniqueName: \"kubernetes.io/projected/30a2e74b-c94d-448a-9d98-a0cc2539be3a-kube-api-access-v6ssz\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:27 crc kubenswrapper[4799]: I1124 08:10:27.008538 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-scripts\") pod \"nova-cell1-cell-mapping-b8cxs\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:27 crc kubenswrapper[4799]: I1124 08:10:27.070199 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:27 crc kubenswrapper[4799]: I1124 08:10:27.204950 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:10:27 crc kubenswrapper[4799]: I1124 08:10:27.205215 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:10:27 crc kubenswrapper[4799]: I1124 08:10:27.499447 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-b8cxs"] Nov 24 08:10:27 crc kubenswrapper[4799]: I1124 08:10:27.836557 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b8cxs" event={"ID":"30a2e74b-c94d-448a-9d98-a0cc2539be3a","Type":"ContainerStarted","Data":"19f55051a854cef73d3e41a738abdda73b0fd616619d9dd47165d5ab2588ccdf"} Nov 24 08:10:27 crc kubenswrapper[4799]: I1124 08:10:27.836619 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b8cxs" event={"ID":"30a2e74b-c94d-448a-9d98-a0cc2539be3a","Type":"ContainerStarted","Data":"86da4b58ad1de34d691d51f22280b650c8918115f47583d30afbb59aac4dc26a"} Nov 24 08:10:27 crc kubenswrapper[4799]: I1124 08:10:27.858193 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-b8cxs" podStartSLOduration=1.8581727749999999 podStartE2EDuration="1.858172775s" podCreationTimestamp="2025-11-24 08:10:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:27.85153167 +0000 UTC m=+4973.507514154" watchObservedRunningTime="2025-11-24 08:10:27.858172775 +0000 UTC m=+4973.514155259" Nov 24 08:10:29 crc kubenswrapper[4799]: I1124 08:10:29.227180 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 08:10:32 crc kubenswrapper[4799]: I1124 08:10:32.204675 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 08:10:32 crc kubenswrapper[4799]: I1124 08:10:32.205546 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 08:10:32 crc kubenswrapper[4799]: I1124 08:10:32.227288 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:10:32 crc kubenswrapper[4799]: I1124 08:10:32.227389 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:10:32 crc kubenswrapper[4799]: I1124 08:10:32.896638 4799 generic.go:334] "Generic (PLEG): container finished" podID="30a2e74b-c94d-448a-9d98-a0cc2539be3a" containerID="19f55051a854cef73d3e41a738abdda73b0fd616619d9dd47165d5ab2588ccdf" exitCode=0 Nov 24 08:10:32 crc kubenswrapper[4799]: I1124 08:10:32.896699 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b8cxs" event={"ID":"30a2e74b-c94d-448a-9d98-a0cc2539be3a","Type":"ContainerDied","Data":"19f55051a854cef73d3e41a738abdda73b0fd616619d9dd47165d5ab2588ccdf"} Nov 24 08:10:33 crc kubenswrapper[4799]: I1124 08:10:33.286109 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.66:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:10:33 crc kubenswrapper[4799]: I1124 08:10:33.368159 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.67:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:10:33 crc kubenswrapper[4799]: I1124 08:10:33.368473 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.66:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:10:33 crc kubenswrapper[4799]: I1124 08:10:33.368492 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.67:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.229174 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.273965 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.304367 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.348571 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-combined-ca-bundle\") pod \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.348916 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-config-data\") pod \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.348984 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-scripts\") pod \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.349062 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6ssz\" (UniqueName: \"kubernetes.io/projected/30a2e74b-c94d-448a-9d98-a0cc2539be3a-kube-api-access-v6ssz\") pod \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\" (UID: \"30a2e74b-c94d-448a-9d98-a0cc2539be3a\") " Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.354750 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a2e74b-c94d-448a-9d98-a0cc2539be3a-kube-api-access-v6ssz" (OuterVolumeSpecName: "kube-api-access-v6ssz") pod "30a2e74b-c94d-448a-9d98-a0cc2539be3a" (UID: "30a2e74b-c94d-448a-9d98-a0cc2539be3a"). InnerVolumeSpecName "kube-api-access-v6ssz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.357516 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-scripts" (OuterVolumeSpecName: "scripts") pod "30a2e74b-c94d-448a-9d98-a0cc2539be3a" (UID: "30a2e74b-c94d-448a-9d98-a0cc2539be3a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.387067 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-config-data" (OuterVolumeSpecName: "config-data") pod "30a2e74b-c94d-448a-9d98-a0cc2539be3a" (UID: "30a2e74b-c94d-448a-9d98-a0cc2539be3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.393355 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30a2e74b-c94d-448a-9d98-a0cc2539be3a" (UID: "30a2e74b-c94d-448a-9d98-a0cc2539be3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.451453 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.451481 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.451490 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a2e74b-c94d-448a-9d98-a0cc2539be3a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.451499 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6ssz\" (UniqueName: \"kubernetes.io/projected/30a2e74b-c94d-448a-9d98-a0cc2539be3a-kube-api-access-v6ssz\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.628522 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:10:34 crc kubenswrapper[4799]: E1124 08:10:34.628805 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.924208 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b8cxs" event={"ID":"30a2e74b-c94d-448a-9d98-a0cc2539be3a","Type":"ContainerDied","Data":"86da4b58ad1de34d691d51f22280b650c8918115f47583d30afbb59aac4dc26a"} Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.924688 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86da4b58ad1de34d691d51f22280b650c8918115f47583d30afbb59aac4dc26a" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.924256 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b8cxs" Nov 24 08:10:34 crc kubenswrapper[4799]: I1124 08:10:34.986403 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 08:10:35 crc kubenswrapper[4799]: I1124 08:10:35.215004 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:35 crc kubenswrapper[4799]: I1124 08:10:35.215275 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerName="nova-api-log" containerID="cri-o://e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b" gracePeriod=30 Nov 24 08:10:35 crc kubenswrapper[4799]: I1124 08:10:35.215429 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerName="nova-api-api" containerID="cri-o://c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a" gracePeriod=30 Nov 24 08:10:35 crc kubenswrapper[4799]: I1124 08:10:35.247277 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:35 crc kubenswrapper[4799]: I1124 08:10:35.247589 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerName="nova-metadata-log" containerID="cri-o://4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296" gracePeriod=30 Nov 24 08:10:35 crc kubenswrapper[4799]: I1124 08:10:35.247712 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerName="nova-metadata-metadata" containerID="cri-o://56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e" gracePeriod=30 Nov 24 08:10:35 crc kubenswrapper[4799]: I1124 08:10:35.419688 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:35 crc kubenswrapper[4799]: I1124 08:10:35.989919 4799 generic.go:334] "Generic (PLEG): container finished" podID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerID="4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296" exitCode=143 Nov 24 08:10:35 crc kubenswrapper[4799]: I1124 08:10:35.989970 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cd4d364-cdfe-4a69-bf54-9858afa7b99e","Type":"ContainerDied","Data":"4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296"} Nov 24 08:10:35 crc kubenswrapper[4799]: I1124 08:10:35.993154 4799 generic.go:334] "Generic (PLEG): container finished" podID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerID="e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b" exitCode=143 Nov 24 08:10:35 crc kubenswrapper[4799]: I1124 08:10:35.993257 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42256f8c-898d-43b7-ac4c-bb98095b5702","Type":"ContainerDied","Data":"e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b"} Nov 24 08:10:37 crc kubenswrapper[4799]: I1124 08:10:37.008221 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6ea30b20-ee93-452f-bf96-9bf7fbcddbe1" containerName="nova-scheduler-scheduler" containerID="cri-o://4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885" gracePeriod=30 Nov 24 08:10:38 crc kubenswrapper[4799]: I1124 08:10:38.988856 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:10:38 crc kubenswrapper[4799]: I1124 08:10:38.991275 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.030669 4799 generic.go:334] "Generic (PLEG): container finished" podID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerID="c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a" exitCode=0 Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.030727 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42256f8c-898d-43b7-ac4c-bb98095b5702","Type":"ContainerDied","Data":"c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a"} Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.030756 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42256f8c-898d-43b7-ac4c-bb98095b5702","Type":"ContainerDied","Data":"72badaebe45f6f77e4bdd4057af93b9a1429a5eede3ca3e59fcd39fb54839d61"} Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.030777 4799 scope.go:117] "RemoveContainer" containerID="c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.030933 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.033729 4799 generic.go:334] "Generic (PLEG): container finished" podID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerID="56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e" exitCode=0 Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.033754 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cd4d364-cdfe-4a69-bf54-9858afa7b99e","Type":"ContainerDied","Data":"56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e"} Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.033773 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cd4d364-cdfe-4a69-bf54-9858afa7b99e","Type":"ContainerDied","Data":"1591b633b17bdcf9ee88610d35bca020850bbe14fd1113db3e9e1ee609a3708e"} Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.033824 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.056022 4799 scope.go:117] "RemoveContainer" containerID="e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.072389 4799 scope.go:117] "RemoveContainer" containerID="c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a" Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.072797 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a\": container with ID starting with c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a not found: ID does not exist" containerID="c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.072830 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a"} err="failed to get container status \"c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a\": rpc error: code = NotFound desc = could not find container \"c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a\": container with ID starting with c0f8ed43d75f0b11caa26ea36bf838f20514369a36f6a6fd62d026feed9ccb6a not found: ID does not exist" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.072867 4799 scope.go:117] "RemoveContainer" containerID="e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b" Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.073104 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b\": container with ID starting with e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b not found: ID does not exist" containerID="e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.073129 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b"} err="failed to get container status \"e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b\": rpc error: code = NotFound desc = could not find container \"e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b\": container with ID starting with e60d3725efc7b5c27008cae7fd16c3be909bbe9eefa71cf021847148d986217b not found: ID does not exist" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.073151 4799 scope.go:117] "RemoveContainer" containerID="56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.091447 4799 scope.go:117] "RemoveContainer" containerID="4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.109371 4799 scope.go:117] "RemoveContainer" containerID="56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e" Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.109705 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e\": container with ID starting with 56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e not found: ID does not exist" containerID="56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.109740 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e"} err="failed to get container status \"56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e\": rpc error: code = NotFound desc = could not find container \"56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e\": container with ID starting with 56360c267dc5e2b5cc3d58945cc70ad6bffc6d535566a7d561e751c4d536ab2e not found: ID does not exist" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.109762 4799 scope.go:117] "RemoveContainer" containerID="4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296" Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.110000 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296\": container with ID starting with 4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296 not found: ID does not exist" containerID="4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.110023 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296"} err="failed to get container status \"4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296\": rpc error: code = NotFound desc = could not find container \"4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296\": container with ID starting with 4ad6694bea4bb99d1b6de613c4180c7a47dfbca7121792a4119d4e24dcd0e296 not found: ID does not exist" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.143405 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-combined-ca-bundle\") pod \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.143452 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-logs\") pod \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.143540 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-config-data\") pod \"42256f8c-898d-43b7-ac4c-bb98095b5702\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.143606 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hb5hw\" (UniqueName: \"kubernetes.io/projected/42256f8c-898d-43b7-ac4c-bb98095b5702-kube-api-access-hb5hw\") pod \"42256f8c-898d-43b7-ac4c-bb98095b5702\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.143659 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42256f8c-898d-43b7-ac4c-bb98095b5702-logs\") pod \"42256f8c-898d-43b7-ac4c-bb98095b5702\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.143697 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-config-data\") pod \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.144391 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-logs" (OuterVolumeSpecName: "logs") pod "6cd4d364-cdfe-4a69-bf54-9858afa7b99e" (UID: "6cd4d364-cdfe-4a69-bf54-9858afa7b99e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.144874 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-combined-ca-bundle\") pod \"42256f8c-898d-43b7-ac4c-bb98095b5702\" (UID: \"42256f8c-898d-43b7-ac4c-bb98095b5702\") " Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.144897 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42256f8c-898d-43b7-ac4c-bb98095b5702-logs" (OuterVolumeSpecName: "logs") pod "42256f8c-898d-43b7-ac4c-bb98095b5702" (UID: "42256f8c-898d-43b7-ac4c-bb98095b5702"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.144904 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvbhs\" (UniqueName: \"kubernetes.io/projected/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-kube-api-access-zvbhs\") pod \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\" (UID: \"6cd4d364-cdfe-4a69-bf54-9858afa7b99e\") " Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.145776 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42256f8c-898d-43b7-ac4c-bb98095b5702-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.145793 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.148913 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-kube-api-access-zvbhs" (OuterVolumeSpecName: "kube-api-access-zvbhs") pod "6cd4d364-cdfe-4a69-bf54-9858afa7b99e" (UID: "6cd4d364-cdfe-4a69-bf54-9858afa7b99e"). InnerVolumeSpecName "kube-api-access-zvbhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.149053 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42256f8c-898d-43b7-ac4c-bb98095b5702-kube-api-access-hb5hw" (OuterVolumeSpecName: "kube-api-access-hb5hw") pod "42256f8c-898d-43b7-ac4c-bb98095b5702" (UID: "42256f8c-898d-43b7-ac4c-bb98095b5702"). InnerVolumeSpecName "kube-api-access-hb5hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.170840 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42256f8c-898d-43b7-ac4c-bb98095b5702" (UID: "42256f8c-898d-43b7-ac4c-bb98095b5702"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.170942 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-config-data" (OuterVolumeSpecName: "config-data") pod "42256f8c-898d-43b7-ac4c-bb98095b5702" (UID: "42256f8c-898d-43b7-ac4c-bb98095b5702"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.171868 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-config-data" (OuterVolumeSpecName: "config-data") pod "6cd4d364-cdfe-4a69-bf54-9858afa7b99e" (UID: "6cd4d364-cdfe-4a69-bf54-9858afa7b99e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.175253 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cd4d364-cdfe-4a69-bf54-9858afa7b99e" (UID: "6cd4d364-cdfe-4a69-bf54-9858afa7b99e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.231007 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.234578 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.236182 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.236267 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="6ea30b20-ee93-452f-bf96-9bf7fbcddbe1" containerName="nova-scheduler-scheduler" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.247137 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvbhs\" (UniqueName: \"kubernetes.io/projected/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-kube-api-access-zvbhs\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.247643 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.247735 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.247801 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hb5hw\" (UniqueName: \"kubernetes.io/projected/42256f8c-898d-43b7-ac4c-bb98095b5702-kube-api-access-hb5hw\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.247882 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cd4d364-cdfe-4a69-bf54-9858afa7b99e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.247950 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42256f8c-898d-43b7-ac4c-bb98095b5702-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.364337 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.374806 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.383666 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.400619 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.421647 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.422215 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerName="nova-metadata-log" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.422244 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerName="nova-metadata-log" Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.422260 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a2e74b-c94d-448a-9d98-a0cc2539be3a" containerName="nova-manage" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.422268 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a2e74b-c94d-448a-9d98-a0cc2539be3a" containerName="nova-manage" Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.422280 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerName="nova-api-log" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.422288 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerName="nova-api-log" Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.422297 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerName="nova-metadata-metadata" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.422305 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerName="nova-metadata-metadata" Nov 24 08:10:39 crc kubenswrapper[4799]: E1124 08:10:39.422348 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerName="nova-api-api" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.422387 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerName="nova-api-api" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.422585 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerName="nova-metadata-metadata" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.422613 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerName="nova-api-api" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.422634 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" containerName="nova-metadata-log" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.422645 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="42256f8c-898d-43b7-ac4c-bb98095b5702" containerName="nova-api-log" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.422655 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a2e74b-c94d-448a-9d98-a0cc2539be3a" containerName="nova-manage" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.424088 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.426095 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.433191 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.435107 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.437056 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.443292 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.452769 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.552103 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-config-data\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.552152 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.552183 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-logs\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.552249 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgsdk\" (UniqueName: \"kubernetes.io/projected/3f6e1cbd-b387-43fd-9c84-523eb571d041-kube-api-access-pgsdk\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.552311 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85s5g\" (UniqueName: \"kubernetes.io/projected/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-kube-api-access-85s5g\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.552350 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f6e1cbd-b387-43fd-9c84-523eb571d041-logs\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.552395 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.552442 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-config-data\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.639743 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42256f8c-898d-43b7-ac4c-bb98095b5702" path="/var/lib/kubelet/pods/42256f8c-898d-43b7-ac4c-bb98095b5702/volumes" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.641282 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cd4d364-cdfe-4a69-bf54-9858afa7b99e" path="/var/lib/kubelet/pods/6cd4d364-cdfe-4a69-bf54-9858afa7b99e/volumes" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.653823 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-config-data\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.653882 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.653907 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-logs\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.653954 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgsdk\" (UniqueName: \"kubernetes.io/projected/3f6e1cbd-b387-43fd-9c84-523eb571d041-kube-api-access-pgsdk\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.653987 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85s5g\" (UniqueName: \"kubernetes.io/projected/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-kube-api-access-85s5g\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.654012 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f6e1cbd-b387-43fd-9c84-523eb571d041-logs\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.654043 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.654070 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-config-data\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.654973 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f6e1cbd-b387-43fd-9c84-523eb571d041-logs\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.655289 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-logs\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.661702 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.662518 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-config-data\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.663231 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-config-data\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.670642 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85s5g\" (UniqueName: \"kubernetes.io/projected/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-kube-api-access-85s5g\") pod \"nova-metadata-0\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " pod="openstack/nova-metadata-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.673464 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgsdk\" (UniqueName: \"kubernetes.io/projected/3f6e1cbd-b387-43fd-9c84-523eb571d041-kube-api-access-pgsdk\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.674001 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.807756 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:10:39 crc kubenswrapper[4799]: I1124 08:10:39.815431 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:10:40 crc kubenswrapper[4799]: I1124 08:10:40.270106 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:10:40 crc kubenswrapper[4799]: I1124 08:10:40.336769 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.059492 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f6e1cbd-b387-43fd-9c84-523eb571d041","Type":"ContainerStarted","Data":"12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958"} Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.059902 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f6e1cbd-b387-43fd-9c84-523eb571d041","Type":"ContainerStarted","Data":"9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad"} Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.059921 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f6e1cbd-b387-43fd-9c84-523eb571d041","Type":"ContainerStarted","Data":"514f6995a94f27b8cba93b8ed2d48ebc3d392cade7a5b3363166147bdce1f6e0"} Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.062052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192","Type":"ContainerStarted","Data":"e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0"} Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.062080 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192","Type":"ContainerStarted","Data":"dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421"} Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.062090 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192","Type":"ContainerStarted","Data":"f3031ac51a9baae4f7d516502477ba3bf0f209003b6a89f01c53819fc8935472"} Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.089165 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.08913815 podStartE2EDuration="2.08913815s" podCreationTimestamp="2025-11-24 08:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:41.082010302 +0000 UTC m=+4986.737992836" watchObservedRunningTime="2025-11-24 08:10:41.08913815 +0000 UTC m=+4986.745120654" Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.114532 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.114509815 podStartE2EDuration="2.114509815s" podCreationTimestamp="2025-11-24 08:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:41.09669292 +0000 UTC m=+4986.752675394" watchObservedRunningTime="2025-11-24 08:10:41.114509815 +0000 UTC m=+4986.770492309" Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.654268 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.790984 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-combined-ca-bundle\") pod \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.791160 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8xg2\" (UniqueName: \"kubernetes.io/projected/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-kube-api-access-q8xg2\") pod \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.791193 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-config-data\") pod \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\" (UID: \"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1\") " Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.796130 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-kube-api-access-q8xg2" (OuterVolumeSpecName: "kube-api-access-q8xg2") pod "6ea30b20-ee93-452f-bf96-9bf7fbcddbe1" (UID: "6ea30b20-ee93-452f-bf96-9bf7fbcddbe1"). InnerVolumeSpecName "kube-api-access-q8xg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:10:41 crc kubenswrapper[4799]: I1124 08:10:41.893910 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8xg2\" (UniqueName: \"kubernetes.io/projected/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-kube-api-access-q8xg2\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.049729 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ea30b20-ee93-452f-bf96-9bf7fbcddbe1" (UID: "6ea30b20-ee93-452f-bf96-9bf7fbcddbe1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.049799 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-config-data" (OuterVolumeSpecName: "config-data") pod "6ea30b20-ee93-452f-bf96-9bf7fbcddbe1" (UID: "6ea30b20-ee93-452f-bf96-9bf7fbcddbe1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.074754 4799 generic.go:334] "Generic (PLEG): container finished" podID="6ea30b20-ee93-452f-bf96-9bf7fbcddbe1" containerID="4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885" exitCode=0 Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.074823 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.074886 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1","Type":"ContainerDied","Data":"4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885"} Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.074961 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6ea30b20-ee93-452f-bf96-9bf7fbcddbe1","Type":"ContainerDied","Data":"62c9003aeeb9316be2a9e8a67febf7d10c75f7e800ebfbd98a4cda8e14c017d5"} Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.074985 4799 scope.go:117] "RemoveContainer" containerID="4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.098305 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.098339 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.130355 4799 scope.go:117] "RemoveContainer" containerID="4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885" Nov 24 08:10:42 crc kubenswrapper[4799]: E1124 08:10:42.130763 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885\": container with ID starting with 4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885 not found: ID does not exist" containerID="4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.130814 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885"} err="failed to get container status \"4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885\": rpc error: code = NotFound desc = could not find container \"4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885\": container with ID starting with 4c19d6296f1edad528f75ef64be1a9f5d8de3491183b20234d1c66c0a1526885 not found: ID does not exist" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.137008 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.154998 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.165037 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:42 crc kubenswrapper[4799]: E1124 08:10:42.165478 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea30b20-ee93-452f-bf96-9bf7fbcddbe1" containerName="nova-scheduler-scheduler" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.165499 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea30b20-ee93-452f-bf96-9bf7fbcddbe1" containerName="nova-scheduler-scheduler" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.165765 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea30b20-ee93-452f-bf96-9bf7fbcddbe1" containerName="nova-scheduler-scheduler" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.166520 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.170644 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.176039 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.301286 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ffvr\" (UniqueName: \"kubernetes.io/projected/2556dfcf-18ac-4b8d-8e92-60784a634311-kube-api-access-8ffvr\") pod \"nova-scheduler-0\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.301394 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.301641 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-config-data\") pod \"nova-scheduler-0\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.403750 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.404141 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-config-data\") pod \"nova-scheduler-0\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.404200 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ffvr\" (UniqueName: \"kubernetes.io/projected/2556dfcf-18ac-4b8d-8e92-60784a634311-kube-api-access-8ffvr\") pod \"nova-scheduler-0\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.408937 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.409837 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-config-data\") pod \"nova-scheduler-0\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.418897 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ffvr\" (UniqueName: \"kubernetes.io/projected/2556dfcf-18ac-4b8d-8e92-60784a634311-kube-api-access-8ffvr\") pod \"nova-scheduler-0\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.483779 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:10:42 crc kubenswrapper[4799]: I1124 08:10:42.935926 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:10:43 crc kubenswrapper[4799]: I1124 08:10:43.084458 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2556dfcf-18ac-4b8d-8e92-60784a634311","Type":"ContainerStarted","Data":"f2c25f4afececd9e4afa9df6dd86dba448b639100e9f8a6711286413f5d40004"} Nov 24 08:10:43 crc kubenswrapper[4799]: I1124 08:10:43.646605 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea30b20-ee93-452f-bf96-9bf7fbcddbe1" path="/var/lib/kubelet/pods/6ea30b20-ee93-452f-bf96-9bf7fbcddbe1/volumes" Nov 24 08:10:44 crc kubenswrapper[4799]: I1124 08:10:44.094010 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2556dfcf-18ac-4b8d-8e92-60784a634311","Type":"ContainerStarted","Data":"8d95a751a9235a693c1cccdb5e380c98d07e547c691e7e4e81213300871c93a3"} Nov 24 08:10:44 crc kubenswrapper[4799]: I1124 08:10:44.112769 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.11275106 podStartE2EDuration="2.11275106s" podCreationTimestamp="2025-11-24 08:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:10:44.107922206 +0000 UTC m=+4989.763904700" watchObservedRunningTime="2025-11-24 08:10:44.11275106 +0000 UTC m=+4989.768733534" Nov 24 08:10:44 crc kubenswrapper[4799]: I1124 08:10:44.816125 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:10:44 crc kubenswrapper[4799]: I1124 08:10:44.817007 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:10:47 crc kubenswrapper[4799]: I1124 08:10:47.484348 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 08:10:49 crc kubenswrapper[4799]: I1124 08:10:49.629164 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:10:49 crc kubenswrapper[4799]: E1124 08:10:49.630310 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:10:49 crc kubenswrapper[4799]: I1124 08:10:49.808599 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:10:49 crc kubenswrapper[4799]: I1124 08:10:49.808655 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:10:49 crc kubenswrapper[4799]: I1124 08:10:49.816154 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 08:10:49 crc kubenswrapper[4799]: I1124 08:10:49.816289 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 08:10:50 crc kubenswrapper[4799]: I1124 08:10:50.973143 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.70:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:10:50 crc kubenswrapper[4799]: I1124 08:10:50.973204 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.71:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:10:50 crc kubenswrapper[4799]: I1124 08:10:50.973209 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.70:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:10:50 crc kubenswrapper[4799]: I1124 08:10:50.973158 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.71:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:10:52 crc kubenswrapper[4799]: I1124 08:10:52.484239 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 08:10:52 crc kubenswrapper[4799]: I1124 08:10:52.513163 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 08:10:53 crc kubenswrapper[4799]: I1124 08:10:53.229332 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 08:10:59 crc kubenswrapper[4799]: I1124 08:10:59.813533 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 08:10:59 crc kubenswrapper[4799]: I1124 08:10:59.814303 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 08:10:59 crc kubenswrapper[4799]: I1124 08:10:59.814700 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 08:10:59 crc kubenswrapper[4799]: I1124 08:10:59.814995 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 08:10:59 crc kubenswrapper[4799]: I1124 08:10:59.817300 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 08:10:59 crc kubenswrapper[4799]: I1124 08:10:59.818619 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 08:10:59 crc kubenswrapper[4799]: I1124 08:10:59.822905 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 08:10:59 crc kubenswrapper[4799]: I1124 08:10:59.824182 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 08:10:59 crc kubenswrapper[4799]: I1124 08:10:59.827871 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.012591 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55c55bdf6f-249sr"] Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.015121 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.025779 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55c55bdf6f-249sr"] Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.192212 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-sb\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.192715 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwzqr\" (UniqueName: \"kubernetes.io/projected/ca758fe2-ced2-4ead-95d7-6302c916d2cd-kube-api-access-zwzqr\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.192860 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-nb\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.192902 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-config\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.192964 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-dns-svc\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.292955 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.293972 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-nb\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.294104 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-config\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.294214 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-dns-svc\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.294343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-sb\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.294491 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwzqr\" (UniqueName: \"kubernetes.io/projected/ca758fe2-ced2-4ead-95d7-6302c916d2cd-kube-api-access-zwzqr\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.295122 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-config\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.295145 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-sb\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.295122 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-dns-svc\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.295784 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-nb\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.335564 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwzqr\" (UniqueName: \"kubernetes.io/projected/ca758fe2-ced2-4ead-95d7-6302c916d2cd-kube-api-access-zwzqr\") pod \"dnsmasq-dns-55c55bdf6f-249sr\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.339067 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.630504 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:11:00 crc kubenswrapper[4799]: E1124 08:11:00.631008 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:11:00 crc kubenswrapper[4799]: I1124 08:11:00.811073 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55c55bdf6f-249sr"] Nov 24 08:11:00 crc kubenswrapper[4799]: W1124 08:11:00.817894 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca758fe2_ced2_4ead_95d7_6302c916d2cd.slice/crio-f62d28591840a62aa0adcc217de17894b6230b042d71c6f239c458974e63df69 WatchSource:0}: Error finding container f62d28591840a62aa0adcc217de17894b6230b042d71c6f239c458974e63df69: Status 404 returned error can't find the container with id f62d28591840a62aa0adcc217de17894b6230b042d71c6f239c458974e63df69 Nov 24 08:11:01 crc kubenswrapper[4799]: I1124 08:11:01.297491 4799 generic.go:334] "Generic (PLEG): container finished" podID="ca758fe2-ced2-4ead-95d7-6302c916d2cd" containerID="b5f774fbaad77232430421a94fc165b44b6e1d361dd436c8b3febb5d75a0826c" exitCode=0 Nov 24 08:11:01 crc kubenswrapper[4799]: I1124 08:11:01.297671 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" event={"ID":"ca758fe2-ced2-4ead-95d7-6302c916d2cd","Type":"ContainerDied","Data":"b5f774fbaad77232430421a94fc165b44b6e1d361dd436c8b3febb5d75a0826c"} Nov 24 08:11:01 crc kubenswrapper[4799]: I1124 08:11:01.297909 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" event={"ID":"ca758fe2-ced2-4ead-95d7-6302c916d2cd","Type":"ContainerStarted","Data":"f62d28591840a62aa0adcc217de17894b6230b042d71c6f239c458974e63df69"} Nov 24 08:11:02 crc kubenswrapper[4799]: I1124 08:11:02.308472 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" event={"ID":"ca758fe2-ced2-4ead-95d7-6302c916d2cd","Type":"ContainerStarted","Data":"cf666b3bb466b2cae0b36d18584fb174550d52af5f00880f3ff1c563665e6c6d"} Nov 24 08:11:02 crc kubenswrapper[4799]: I1124 08:11:02.337668 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" podStartSLOduration=3.33764726 podStartE2EDuration="3.33764726s" podCreationTimestamp="2025-11-24 08:10:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:11:02.332900758 +0000 UTC m=+5007.988883242" watchObservedRunningTime="2025-11-24 08:11:02.33764726 +0000 UTC m=+5007.993629734" Nov 24 08:11:03 crc kubenswrapper[4799]: I1124 08:11:03.319103 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:10 crc kubenswrapper[4799]: I1124 08:11:10.341158 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:11:10 crc kubenswrapper[4799]: I1124 08:11:10.446785 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cb57d6b9-bmvcf"] Nov 24 08:11:10 crc kubenswrapper[4799]: I1124 08:11:10.447059 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" podUID="1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" containerName="dnsmasq-dns" containerID="cri-o://10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14" gracePeriod=10 Nov 24 08:11:10 crc kubenswrapper[4799]: I1124 08:11:10.918214 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:11:10 crc kubenswrapper[4799]: I1124 08:11:10.998776 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-nb\") pod \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " Nov 24 08:11:10 crc kubenswrapper[4799]: I1124 08:11:10.998828 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-sb\") pod \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:10.998917 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-config\") pod \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:10.998995 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-dns-svc\") pod \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:10.999109 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zktlg\" (UniqueName: \"kubernetes.io/projected/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-kube-api-access-zktlg\") pod \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\" (UID: \"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090\") " Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.005292 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-kube-api-access-zktlg" (OuterVolumeSpecName: "kube-api-access-zktlg") pod "1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" (UID: "1e27a52f-03e1-4eea-a5cd-3c2a43c0f090"). InnerVolumeSpecName "kube-api-access-zktlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.046668 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" (UID: "1e27a52f-03e1-4eea-a5cd-3c2a43c0f090"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.046685 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" (UID: "1e27a52f-03e1-4eea-a5cd-3c2a43c0f090"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.060408 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" (UID: "1e27a52f-03e1-4eea-a5cd-3c2a43c0f090"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.065152 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-config" (OuterVolumeSpecName: "config") pod "1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" (UID: "1e27a52f-03e1-4eea-a5cd-3c2a43c0f090"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.101559 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.101825 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zktlg\" (UniqueName: \"kubernetes.io/projected/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-kube-api-access-zktlg\") on node \"crc\" DevicePath \"\"" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.101845 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.101860 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.101896 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.417569 4799 generic.go:334] "Generic (PLEG): container finished" podID="1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" containerID="10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14" exitCode=0 Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.417634 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" event={"ID":"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090","Type":"ContainerDied","Data":"10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14"} Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.417681 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" event={"ID":"1e27a52f-03e1-4eea-a5cd-3c2a43c0f090","Type":"ContainerDied","Data":"15b3c6b09e17acb9871512d866c0cbe7ea94c310b2e315f0429739f2312330a9"} Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.417709 4799 scope.go:117] "RemoveContainer" containerID="10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.417701 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cb57d6b9-bmvcf" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.457803 4799 scope.go:117] "RemoveContainer" containerID="71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.486190 4799 scope.go:117] "RemoveContainer" containerID="10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14" Nov 24 08:11:11 crc kubenswrapper[4799]: E1124 08:11:11.486722 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14\": container with ID starting with 10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14 not found: ID does not exist" containerID="10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.486807 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14"} err="failed to get container status \"10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14\": rpc error: code = NotFound desc = could not find container \"10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14\": container with ID starting with 10b53e41d8037fd3bdb47d6fb5c0f137fac77d890b863a9bc8e73d0d945ecc14 not found: ID does not exist" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.486847 4799 scope.go:117] "RemoveContainer" containerID="71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf" Nov 24 08:11:11 crc kubenswrapper[4799]: E1124 08:11:11.487719 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf\": container with ID starting with 71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf not found: ID does not exist" containerID="71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.488010 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf"} err="failed to get container status \"71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf\": rpc error: code = NotFound desc = could not find container \"71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf\": container with ID starting with 71fee4a9072ba4f34795c1620f161914e70f591f3e8dccfc3a04683047141dbf not found: ID does not exist" Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.497107 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cb57d6b9-bmvcf"] Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.507142 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cb57d6b9-bmvcf"] Nov 24 08:11:11 crc kubenswrapper[4799]: I1124 08:11:11.641721 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" path="/var/lib/kubelet/pods/1e27a52f-03e1-4eea-a5cd-3c2a43c0f090/volumes" Nov 24 08:11:12 crc kubenswrapper[4799]: I1124 08:11:12.628167 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:11:12 crc kubenswrapper[4799]: E1124 08:11:12.628621 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.020100 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-nzj9x"] Nov 24 08:11:14 crc kubenswrapper[4799]: E1124 08:11:14.021162 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" containerName="init" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.021182 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" containerName="init" Nov 24 08:11:14 crc kubenswrapper[4799]: E1124 08:11:14.021225 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" containerName="dnsmasq-dns" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.021233 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" containerName="dnsmasq-dns" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.021439 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e27a52f-03e1-4eea-a5cd-3c2a43c0f090" containerName="dnsmasq-dns" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.022238 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nzj9x" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.032136 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-nzj9x"] Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.048662 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4d6w\" (UniqueName: \"kubernetes.io/projected/caf0c85d-c501-4dde-af9a-c7d5aa13c619-kube-api-access-c4d6w\") pod \"cinder-db-create-nzj9x\" (UID: \"caf0c85d-c501-4dde-af9a-c7d5aa13c619\") " pod="openstack/cinder-db-create-nzj9x" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.048731 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caf0c85d-c501-4dde-af9a-c7d5aa13c619-operator-scripts\") pod \"cinder-db-create-nzj9x\" (UID: \"caf0c85d-c501-4dde-af9a-c7d5aa13c619\") " pod="openstack/cinder-db-create-nzj9x" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.121081 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f4e6-account-create-jtxfl"] Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.122248 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f4e6-account-create-jtxfl" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.124483 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.132031 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f4e6-account-create-jtxfl"] Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.149866 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2504a900-40b9-4f68-8a13-903dac0dd3d6-operator-scripts\") pod \"cinder-f4e6-account-create-jtxfl\" (UID: \"2504a900-40b9-4f68-8a13-903dac0dd3d6\") " pod="openstack/cinder-f4e6-account-create-jtxfl" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.149978 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4d6w\" (UniqueName: \"kubernetes.io/projected/caf0c85d-c501-4dde-af9a-c7d5aa13c619-kube-api-access-c4d6w\") pod \"cinder-db-create-nzj9x\" (UID: \"caf0c85d-c501-4dde-af9a-c7d5aa13c619\") " pod="openstack/cinder-db-create-nzj9x" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.150023 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caf0c85d-c501-4dde-af9a-c7d5aa13c619-operator-scripts\") pod \"cinder-db-create-nzj9x\" (UID: \"caf0c85d-c501-4dde-af9a-c7d5aa13c619\") " pod="openstack/cinder-db-create-nzj9x" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.150077 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dqdp\" (UniqueName: \"kubernetes.io/projected/2504a900-40b9-4f68-8a13-903dac0dd3d6-kube-api-access-7dqdp\") pod \"cinder-f4e6-account-create-jtxfl\" (UID: \"2504a900-40b9-4f68-8a13-903dac0dd3d6\") " pod="openstack/cinder-f4e6-account-create-jtxfl" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.150819 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caf0c85d-c501-4dde-af9a-c7d5aa13c619-operator-scripts\") pod \"cinder-db-create-nzj9x\" (UID: \"caf0c85d-c501-4dde-af9a-c7d5aa13c619\") " pod="openstack/cinder-db-create-nzj9x" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.181548 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4d6w\" (UniqueName: \"kubernetes.io/projected/caf0c85d-c501-4dde-af9a-c7d5aa13c619-kube-api-access-c4d6w\") pod \"cinder-db-create-nzj9x\" (UID: \"caf0c85d-c501-4dde-af9a-c7d5aa13c619\") " pod="openstack/cinder-db-create-nzj9x" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.252174 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2504a900-40b9-4f68-8a13-903dac0dd3d6-operator-scripts\") pod \"cinder-f4e6-account-create-jtxfl\" (UID: \"2504a900-40b9-4f68-8a13-903dac0dd3d6\") " pod="openstack/cinder-f4e6-account-create-jtxfl" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.252349 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dqdp\" (UniqueName: \"kubernetes.io/projected/2504a900-40b9-4f68-8a13-903dac0dd3d6-kube-api-access-7dqdp\") pod \"cinder-f4e6-account-create-jtxfl\" (UID: \"2504a900-40b9-4f68-8a13-903dac0dd3d6\") " pod="openstack/cinder-f4e6-account-create-jtxfl" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.252976 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2504a900-40b9-4f68-8a13-903dac0dd3d6-operator-scripts\") pod \"cinder-f4e6-account-create-jtxfl\" (UID: \"2504a900-40b9-4f68-8a13-903dac0dd3d6\") " pod="openstack/cinder-f4e6-account-create-jtxfl" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.268079 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dqdp\" (UniqueName: \"kubernetes.io/projected/2504a900-40b9-4f68-8a13-903dac0dd3d6-kube-api-access-7dqdp\") pod \"cinder-f4e6-account-create-jtxfl\" (UID: \"2504a900-40b9-4f68-8a13-903dac0dd3d6\") " pod="openstack/cinder-f4e6-account-create-jtxfl" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.346522 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nzj9x" Nov 24 08:11:14 crc kubenswrapper[4799]: I1124 08:11:14.440604 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f4e6-account-create-jtxfl" Nov 24 08:11:15 crc kubenswrapper[4799]: I1124 08:11:14.857870 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-nzj9x"] Nov 24 08:11:15 crc kubenswrapper[4799]: I1124 08:11:14.958079 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f4e6-account-create-jtxfl"] Nov 24 08:11:15 crc kubenswrapper[4799]: W1124 08:11:14.977380 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2504a900_40b9_4f68_8a13_903dac0dd3d6.slice/crio-b5ddf60fc828f3ca5e50fffde793aba9d67ba766568b4e48046e270b77330c91 WatchSource:0}: Error finding container b5ddf60fc828f3ca5e50fffde793aba9d67ba766568b4e48046e270b77330c91: Status 404 returned error can't find the container with id b5ddf60fc828f3ca5e50fffde793aba9d67ba766568b4e48046e270b77330c91 Nov 24 08:11:15 crc kubenswrapper[4799]: I1124 08:11:15.463156 4799 generic.go:334] "Generic (PLEG): container finished" podID="caf0c85d-c501-4dde-af9a-c7d5aa13c619" containerID="458c646435fa2aed0777b74c9c4698dc45a17fc9d18d985ee8a69a253f0768df" exitCode=0 Nov 24 08:11:15 crc kubenswrapper[4799]: I1124 08:11:15.463282 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nzj9x" event={"ID":"caf0c85d-c501-4dde-af9a-c7d5aa13c619","Type":"ContainerDied","Data":"458c646435fa2aed0777b74c9c4698dc45a17fc9d18d985ee8a69a253f0768df"} Nov 24 08:11:15 crc kubenswrapper[4799]: I1124 08:11:15.463366 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nzj9x" event={"ID":"caf0c85d-c501-4dde-af9a-c7d5aa13c619","Type":"ContainerStarted","Data":"bbd1e88ac71fa4509c578033777011af9d75a30c5892f4a4f4f01cd390f726c2"} Nov 24 08:11:15 crc kubenswrapper[4799]: I1124 08:11:15.466252 4799 generic.go:334] "Generic (PLEG): container finished" podID="2504a900-40b9-4f68-8a13-903dac0dd3d6" containerID="477a30ce6f0d990499994aefcea2d8a135d3fa9ccf6de85de7c5144a5c393a5a" exitCode=0 Nov 24 08:11:15 crc kubenswrapper[4799]: I1124 08:11:15.466315 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f4e6-account-create-jtxfl" event={"ID":"2504a900-40b9-4f68-8a13-903dac0dd3d6","Type":"ContainerDied","Data":"477a30ce6f0d990499994aefcea2d8a135d3fa9ccf6de85de7c5144a5c393a5a"} Nov 24 08:11:15 crc kubenswrapper[4799]: I1124 08:11:15.466353 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f4e6-account-create-jtxfl" event={"ID":"2504a900-40b9-4f68-8a13-903dac0dd3d6","Type":"ContainerStarted","Data":"b5ddf60fc828f3ca5e50fffde793aba9d67ba766568b4e48046e270b77330c91"} Nov 24 08:11:16 crc kubenswrapper[4799]: I1124 08:11:16.907579 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nzj9x" Nov 24 08:11:16 crc kubenswrapper[4799]: I1124 08:11:16.916662 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f4e6-account-create-jtxfl" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.043191 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dqdp\" (UniqueName: \"kubernetes.io/projected/2504a900-40b9-4f68-8a13-903dac0dd3d6-kube-api-access-7dqdp\") pod \"2504a900-40b9-4f68-8a13-903dac0dd3d6\" (UID: \"2504a900-40b9-4f68-8a13-903dac0dd3d6\") " Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.043635 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2504a900-40b9-4f68-8a13-903dac0dd3d6-operator-scripts\") pod \"2504a900-40b9-4f68-8a13-903dac0dd3d6\" (UID: \"2504a900-40b9-4f68-8a13-903dac0dd3d6\") " Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.043676 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caf0c85d-c501-4dde-af9a-c7d5aa13c619-operator-scripts\") pod \"caf0c85d-c501-4dde-af9a-c7d5aa13c619\" (UID: \"caf0c85d-c501-4dde-af9a-c7d5aa13c619\") " Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.043706 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4d6w\" (UniqueName: \"kubernetes.io/projected/caf0c85d-c501-4dde-af9a-c7d5aa13c619-kube-api-access-c4d6w\") pod \"caf0c85d-c501-4dde-af9a-c7d5aa13c619\" (UID: \"caf0c85d-c501-4dde-af9a-c7d5aa13c619\") " Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.044271 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2504a900-40b9-4f68-8a13-903dac0dd3d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2504a900-40b9-4f68-8a13-903dac0dd3d6" (UID: "2504a900-40b9-4f68-8a13-903dac0dd3d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.044675 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caf0c85d-c501-4dde-af9a-c7d5aa13c619-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "caf0c85d-c501-4dde-af9a-c7d5aa13c619" (UID: "caf0c85d-c501-4dde-af9a-c7d5aa13c619"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.048483 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2504a900-40b9-4f68-8a13-903dac0dd3d6-kube-api-access-7dqdp" (OuterVolumeSpecName: "kube-api-access-7dqdp") pod "2504a900-40b9-4f68-8a13-903dac0dd3d6" (UID: "2504a900-40b9-4f68-8a13-903dac0dd3d6"). InnerVolumeSpecName "kube-api-access-7dqdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.049886 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caf0c85d-c501-4dde-af9a-c7d5aa13c619-kube-api-access-c4d6w" (OuterVolumeSpecName: "kube-api-access-c4d6w") pod "caf0c85d-c501-4dde-af9a-c7d5aa13c619" (UID: "caf0c85d-c501-4dde-af9a-c7d5aa13c619"). InnerVolumeSpecName "kube-api-access-c4d6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.147508 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dqdp\" (UniqueName: \"kubernetes.io/projected/2504a900-40b9-4f68-8a13-903dac0dd3d6-kube-api-access-7dqdp\") on node \"crc\" DevicePath \"\"" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.147550 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2504a900-40b9-4f68-8a13-903dac0dd3d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.147563 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caf0c85d-c501-4dde-af9a-c7d5aa13c619-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.147573 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4d6w\" (UniqueName: \"kubernetes.io/projected/caf0c85d-c501-4dde-af9a-c7d5aa13c619-kube-api-access-c4d6w\") on node \"crc\" DevicePath \"\"" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.487031 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nzj9x" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.487041 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nzj9x" event={"ID":"caf0c85d-c501-4dde-af9a-c7d5aa13c619","Type":"ContainerDied","Data":"bbd1e88ac71fa4509c578033777011af9d75a30c5892f4a4f4f01cd390f726c2"} Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.487106 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbd1e88ac71fa4509c578033777011af9d75a30c5892f4a4f4f01cd390f726c2" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.492914 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f4e6-account-create-jtxfl" event={"ID":"2504a900-40b9-4f68-8a13-903dac0dd3d6","Type":"ContainerDied","Data":"b5ddf60fc828f3ca5e50fffde793aba9d67ba766568b4e48046e270b77330c91"} Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.492982 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5ddf60fc828f3ca5e50fffde793aba9d67ba766568b4e48046e270b77330c91" Nov 24 08:11:17 crc kubenswrapper[4799]: I1124 08:11:17.493057 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f4e6-account-create-jtxfl" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.310813 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-r7q8c"] Nov 24 08:11:19 crc kubenswrapper[4799]: E1124 08:11:19.311439 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf0c85d-c501-4dde-af9a-c7d5aa13c619" containerName="mariadb-database-create" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.311456 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf0c85d-c501-4dde-af9a-c7d5aa13c619" containerName="mariadb-database-create" Nov 24 08:11:19 crc kubenswrapper[4799]: E1124 08:11:19.311474 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2504a900-40b9-4f68-8a13-903dac0dd3d6" containerName="mariadb-account-create" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.311481 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2504a900-40b9-4f68-8a13-903dac0dd3d6" containerName="mariadb-account-create" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.311665 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2504a900-40b9-4f68-8a13-903dac0dd3d6" containerName="mariadb-account-create" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.311682 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf0c85d-c501-4dde-af9a-c7d5aa13c619" containerName="mariadb-database-create" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.312422 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.315710 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.315977 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.316714 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7nmkt" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.325877 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-r7q8c"] Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.388728 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-db-sync-config-data\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.388805 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-scripts\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.389195 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-combined-ca-bundle\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.389296 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-config-data\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.389501 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5lqt\" (UniqueName: \"kubernetes.io/projected/45a30475-100a-4388-9022-0f1f18eacb82-kube-api-access-g5lqt\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.389834 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/45a30475-100a-4388-9022-0f1f18eacb82-etc-machine-id\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.493185 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-db-sync-config-data\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.493249 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-scripts\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.493328 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-combined-ca-bundle\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.493348 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-config-data\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.493384 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5lqt\" (UniqueName: \"kubernetes.io/projected/45a30475-100a-4388-9022-0f1f18eacb82-kube-api-access-g5lqt\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.493440 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/45a30475-100a-4388-9022-0f1f18eacb82-etc-machine-id\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.493592 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/45a30475-100a-4388-9022-0f1f18eacb82-etc-machine-id\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.514060 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-combined-ca-bundle\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.514340 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-scripts\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.515518 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-db-sync-config-data\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.516508 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-config-data\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.520452 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5lqt\" (UniqueName: \"kubernetes.io/projected/45a30475-100a-4388-9022-0f1f18eacb82-kube-api-access-g5lqt\") pod \"cinder-db-sync-r7q8c\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:19 crc kubenswrapper[4799]: I1124 08:11:19.638975 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:11:20 crc kubenswrapper[4799]: W1124 08:11:20.163355 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45a30475_100a_4388_9022_0f1f18eacb82.slice/crio-5fd74ecd350a24a003fd0b40e864a2764187082a6f909b444f4aa1cad3564109 WatchSource:0}: Error finding container 5fd74ecd350a24a003fd0b40e864a2764187082a6f909b444f4aa1cad3564109: Status 404 returned error can't find the container with id 5fd74ecd350a24a003fd0b40e864a2764187082a6f909b444f4aa1cad3564109 Nov 24 08:11:20 crc kubenswrapper[4799]: I1124 08:11:20.165775 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-r7q8c"] Nov 24 08:11:20 crc kubenswrapper[4799]: I1124 08:11:20.529669 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r7q8c" event={"ID":"45a30475-100a-4388-9022-0f1f18eacb82","Type":"ContainerStarted","Data":"5fd74ecd350a24a003fd0b40e864a2764187082a6f909b444f4aa1cad3564109"} Nov 24 08:11:26 crc kubenswrapper[4799]: I1124 08:11:26.628631 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:11:26 crc kubenswrapper[4799]: E1124 08:11:26.629619 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:11:38 crc kubenswrapper[4799]: I1124 08:11:38.628471 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:11:38 crc kubenswrapper[4799]: E1124 08:11:38.629348 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:11:40 crc kubenswrapper[4799]: E1124 08:11:40.004980 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:94a34fda2d142cfe9e3097b1d1bd6839" Nov 24 08:11:40 crc kubenswrapper[4799]: E1124 08:11:40.005043 4799 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:94a34fda2d142cfe9e3097b1d1bd6839" Nov 24 08:11:40 crc kubenswrapper[4799]: E1124 08:11:40.005169 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:94a34fda2d142cfe9e3097b1d1bd6839,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g5lqt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-r7q8c_openstack(45a30475-100a-4388-9022-0f1f18eacb82): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:11:40 crc kubenswrapper[4799]: E1124 08:11:40.006356 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-r7q8c" podUID="45a30475-100a-4388-9022-0f1f18eacb82" Nov 24 08:11:40 crc kubenswrapper[4799]: E1124 08:11:40.734040 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:94a34fda2d142cfe9e3097b1d1bd6839\\\"\"" pod="openstack/cinder-db-sync-r7q8c" podUID="45a30475-100a-4388-9022-0f1f18eacb82" Nov 24 08:11:49 crc kubenswrapper[4799]: I1124 08:11:49.628884 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:11:49 crc kubenswrapper[4799]: E1124 08:11:49.629603 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:11:56 crc kubenswrapper[4799]: I1124 08:11:56.884064 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r7q8c" event={"ID":"45a30475-100a-4388-9022-0f1f18eacb82","Type":"ContainerStarted","Data":"72714e01cd15036817855eabdf3bdb17d708dc9946e6d3c97c894c55df400e76"} Nov 24 08:11:56 crc kubenswrapper[4799]: I1124 08:11:56.901184 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-r7q8c" podStartSLOduration=2.198502272 podStartE2EDuration="37.901166849s" podCreationTimestamp="2025-11-24 08:11:19 +0000 UTC" firstStartedPulling="2025-11-24 08:11:20.16772778 +0000 UTC m=+5025.823710294" lastFinishedPulling="2025-11-24 08:11:55.870392397 +0000 UTC m=+5061.526374871" observedRunningTime="2025-11-24 08:11:56.897749032 +0000 UTC m=+5062.553731506" watchObservedRunningTime="2025-11-24 08:11:56.901166849 +0000 UTC m=+5062.557149323" Nov 24 08:11:59 crc kubenswrapper[4799]: I1124 08:11:59.940910 4799 generic.go:334] "Generic (PLEG): container finished" podID="45a30475-100a-4388-9022-0f1f18eacb82" containerID="72714e01cd15036817855eabdf3bdb17d708dc9946e6d3c97c894c55df400e76" exitCode=0 Nov 24 08:11:59 crc kubenswrapper[4799]: I1124 08:11:59.941010 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r7q8c" event={"ID":"45a30475-100a-4388-9022-0f1f18eacb82","Type":"ContainerDied","Data":"72714e01cd15036817855eabdf3bdb17d708dc9946e6d3c97c894c55df400e76"} Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.310583 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.406482 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/45a30475-100a-4388-9022-0f1f18eacb82-etc-machine-id\") pod \"45a30475-100a-4388-9022-0f1f18eacb82\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.406590 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45a30475-100a-4388-9022-0f1f18eacb82-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "45a30475-100a-4388-9022-0f1f18eacb82" (UID: "45a30475-100a-4388-9022-0f1f18eacb82"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.406605 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-combined-ca-bundle\") pod \"45a30475-100a-4388-9022-0f1f18eacb82\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.406697 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-db-sync-config-data\") pod \"45a30475-100a-4388-9022-0f1f18eacb82\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.406790 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-scripts\") pod \"45a30475-100a-4388-9022-0f1f18eacb82\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.406882 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5lqt\" (UniqueName: \"kubernetes.io/projected/45a30475-100a-4388-9022-0f1f18eacb82-kube-api-access-g5lqt\") pod \"45a30475-100a-4388-9022-0f1f18eacb82\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.406946 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-config-data\") pod \"45a30475-100a-4388-9022-0f1f18eacb82\" (UID: \"45a30475-100a-4388-9022-0f1f18eacb82\") " Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.407721 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/45a30475-100a-4388-9022-0f1f18eacb82-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.412475 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "45a30475-100a-4388-9022-0f1f18eacb82" (UID: "45a30475-100a-4388-9022-0f1f18eacb82"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.412856 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-scripts" (OuterVolumeSpecName: "scripts") pod "45a30475-100a-4388-9022-0f1f18eacb82" (UID: "45a30475-100a-4388-9022-0f1f18eacb82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.422592 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45a30475-100a-4388-9022-0f1f18eacb82-kube-api-access-g5lqt" (OuterVolumeSpecName: "kube-api-access-g5lqt") pod "45a30475-100a-4388-9022-0f1f18eacb82" (UID: "45a30475-100a-4388-9022-0f1f18eacb82"). InnerVolumeSpecName "kube-api-access-g5lqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.437233 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45a30475-100a-4388-9022-0f1f18eacb82" (UID: "45a30475-100a-4388-9022-0f1f18eacb82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.457843 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-config-data" (OuterVolumeSpecName: "config-data") pod "45a30475-100a-4388-9022-0f1f18eacb82" (UID: "45a30475-100a-4388-9022-0f1f18eacb82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.509700 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.509741 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.509755 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.509766 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5lqt\" (UniqueName: \"kubernetes.io/projected/45a30475-100a-4388-9022-0f1f18eacb82-kube-api-access-g5lqt\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.509783 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45a30475-100a-4388-9022-0f1f18eacb82-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.630313 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:12:01 crc kubenswrapper[4799]: E1124 08:12:01.630548 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.968335 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r7q8c" event={"ID":"45a30475-100a-4388-9022-0f1f18eacb82","Type":"ContainerDied","Data":"5fd74ecd350a24a003fd0b40e864a2764187082a6f909b444f4aa1cad3564109"} Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.968706 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fd74ecd350a24a003fd0b40e864a2764187082a6f909b444f4aa1cad3564109" Nov 24 08:12:01 crc kubenswrapper[4799]: I1124 08:12:01.968595 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r7q8c" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.299982 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b86c965c-lhnzm"] Nov 24 08:12:02 crc kubenswrapper[4799]: E1124 08:12:02.300425 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45a30475-100a-4388-9022-0f1f18eacb82" containerName="cinder-db-sync" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.300443 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="45a30475-100a-4388-9022-0f1f18eacb82" containerName="cinder-db-sync" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.300712 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="45a30475-100a-4388-9022-0f1f18eacb82" containerName="cinder-db-sync" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.301883 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.362968 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b86c965c-lhnzm"] Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.430125 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.430207 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-dns-svc\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.430278 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-config\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.430302 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6jk2\" (UniqueName: \"kubernetes.io/projected/27b56ac5-d465-4341-a689-bd7c4dc5438c-kube-api-access-x6jk2\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.430369 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.531446 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.531502 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-dns-svc\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.531552 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-config\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.531571 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6jk2\" (UniqueName: \"kubernetes.io/projected/27b56ac5-d465-4341-a689-bd7c4dc5438c-kube-api-access-x6jk2\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.531593 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.532431 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.532473 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.532523 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-config\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.532587 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-dns-svc\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.582216 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6jk2\" (UniqueName: \"kubernetes.io/projected/27b56ac5-d465-4341-a689-bd7c4dc5438c-kube-api-access-x6jk2\") pod \"dnsmasq-dns-67b86c965c-lhnzm\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.618319 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.643998 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.648825 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.656179 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.656747 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.656496 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.657695 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7nmkt" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.690995 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.742669 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f05f65a-f0e1-4645-8085-5614e23dc198-logs\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.743007 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data-custom\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.743161 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.743342 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjxk7\" (UniqueName: \"kubernetes.io/projected/7f05f65a-f0e1-4645-8085-5614e23dc198-kube-api-access-kjxk7\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.743473 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.743578 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f05f65a-f0e1-4645-8085-5614e23dc198-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.743747 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-scripts\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.844772 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjxk7\" (UniqueName: \"kubernetes.io/projected/7f05f65a-f0e1-4645-8085-5614e23dc198-kube-api-access-kjxk7\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.844822 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.844865 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f05f65a-f0e1-4645-8085-5614e23dc198-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.844916 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-scripts\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.844947 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f05f65a-f0e1-4645-8085-5614e23dc198-logs\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.844964 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data-custom\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.844984 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.847067 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f05f65a-f0e1-4645-8085-5614e23dc198-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.853445 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f05f65a-f0e1-4645-8085-5614e23dc198-logs\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.861397 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.861805 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-scripts\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.869627 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjxk7\" (UniqueName: \"kubernetes.io/projected/7f05f65a-f0e1-4645-8085-5614e23dc198-kube-api-access-kjxk7\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.869683 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:02 crc kubenswrapper[4799]: I1124 08:12:02.873380 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data-custom\") pod \"cinder-api-0\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " pod="openstack/cinder-api-0" Nov 24 08:12:03 crc kubenswrapper[4799]: I1124 08:12:03.048380 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:12:03 crc kubenswrapper[4799]: I1124 08:12:03.167839 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b86c965c-lhnzm"] Nov 24 08:12:03 crc kubenswrapper[4799]: W1124 08:12:03.179104 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27b56ac5_d465_4341_a689_bd7c4dc5438c.slice/crio-104af59ae6f184eb771eeaf90760ac5398d8db3830602a0d9c2e8d4e1e18e5cb WatchSource:0}: Error finding container 104af59ae6f184eb771eeaf90760ac5398d8db3830602a0d9c2e8d4e1e18e5cb: Status 404 returned error can't find the container with id 104af59ae6f184eb771eeaf90760ac5398d8db3830602a0d9c2e8d4e1e18e5cb Nov 24 08:12:03 crc kubenswrapper[4799]: I1124 08:12:03.361841 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:12:03 crc kubenswrapper[4799]: W1124 08:12:03.370108 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f05f65a_f0e1_4645_8085_5614e23dc198.slice/crio-d80b4683eb83a57e5ad16c4738a278bca3683a462c6107f11dc26c2500ff5934 WatchSource:0}: Error finding container d80b4683eb83a57e5ad16c4738a278bca3683a462c6107f11dc26c2500ff5934: Status 404 returned error can't find the container with id d80b4683eb83a57e5ad16c4738a278bca3683a462c6107f11dc26c2500ff5934 Nov 24 08:12:03 crc kubenswrapper[4799]: I1124 08:12:03.988259 4799 generic.go:334] "Generic (PLEG): container finished" podID="27b56ac5-d465-4341-a689-bd7c4dc5438c" containerID="5ad23e59167853cdf35ee4240b3de589d0fd11b88d2a7f5e8042bdbad847ddd1" exitCode=0 Nov 24 08:12:03 crc kubenswrapper[4799]: I1124 08:12:03.988368 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" event={"ID":"27b56ac5-d465-4341-a689-bd7c4dc5438c","Type":"ContainerDied","Data":"5ad23e59167853cdf35ee4240b3de589d0fd11b88d2a7f5e8042bdbad847ddd1"} Nov 24 08:12:03 crc kubenswrapper[4799]: I1124 08:12:03.988757 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" event={"ID":"27b56ac5-d465-4341-a689-bd7c4dc5438c","Type":"ContainerStarted","Data":"104af59ae6f184eb771eeaf90760ac5398d8db3830602a0d9c2e8d4e1e18e5cb"} Nov 24 08:12:03 crc kubenswrapper[4799]: I1124 08:12:03.990451 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f05f65a-f0e1-4645-8085-5614e23dc198","Type":"ContainerStarted","Data":"d80b4683eb83a57e5ad16c4738a278bca3683a462c6107f11dc26c2500ff5934"} Nov 24 08:12:05 crc kubenswrapper[4799]: I1124 08:12:05.000480 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f05f65a-f0e1-4645-8085-5614e23dc198","Type":"ContainerStarted","Data":"2a794cfe903d6ddfe89e23172c37df55f73ab3305a535a8ea7f5d0413d7a414b"} Nov 24 08:12:05 crc kubenswrapper[4799]: I1124 08:12:05.001147 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 08:12:05 crc kubenswrapper[4799]: I1124 08:12:05.001170 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f05f65a-f0e1-4645-8085-5614e23dc198","Type":"ContainerStarted","Data":"db78d862d8e4edd6078acf7c227757b2f6bd51963574eecb371a2dfe75e3b3bc"} Nov 24 08:12:05 crc kubenswrapper[4799]: I1124 08:12:05.002082 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" event={"ID":"27b56ac5-d465-4341-a689-bd7c4dc5438c","Type":"ContainerStarted","Data":"2c247ff0d900ac39728dae153b1a08256c5ba8196fad15da6205848517a7f697"} Nov 24 08:12:05 crc kubenswrapper[4799]: I1124 08:12:05.003261 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:05 crc kubenswrapper[4799]: I1124 08:12:05.019196 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.019177594 podStartE2EDuration="3.019177594s" podCreationTimestamp="2025-11-24 08:12:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:12:05.014524702 +0000 UTC m=+5070.670507176" watchObservedRunningTime="2025-11-24 08:12:05.019177594 +0000 UTC m=+5070.675160068" Nov 24 08:12:05 crc kubenswrapper[4799]: I1124 08:12:05.037291 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" podStartSLOduration=3.037273658 podStartE2EDuration="3.037273658s" podCreationTimestamp="2025-11-24 08:12:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:12:05.0317141 +0000 UTC m=+5070.687696574" watchObservedRunningTime="2025-11-24 08:12:05.037273658 +0000 UTC m=+5070.693256132" Nov 24 08:12:12 crc kubenswrapper[4799]: I1124 08:12:12.621099 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:12:12 crc kubenswrapper[4799]: I1124 08:12:12.628572 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:12:12 crc kubenswrapper[4799]: E1124 08:12:12.628983 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:12:12 crc kubenswrapper[4799]: I1124 08:12:12.690742 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55c55bdf6f-249sr"] Nov 24 08:12:12 crc kubenswrapper[4799]: I1124 08:12:12.690995 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" podUID="ca758fe2-ced2-4ead-95d7-6302c916d2cd" containerName="dnsmasq-dns" containerID="cri-o://cf666b3bb466b2cae0b36d18584fb174550d52af5f00880f3ff1c563665e6c6d" gracePeriod=10 Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.107415 4799 generic.go:334] "Generic (PLEG): container finished" podID="ca758fe2-ced2-4ead-95d7-6302c916d2cd" containerID="cf666b3bb466b2cae0b36d18584fb174550d52af5f00880f3ff1c563665e6c6d" exitCode=0 Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.107600 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" event={"ID":"ca758fe2-ced2-4ead-95d7-6302c916d2cd","Type":"ContainerDied","Data":"cf666b3bb466b2cae0b36d18584fb174550d52af5f00880f3ff1c563665e6c6d"} Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.279735 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.344976 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-nb\") pod \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.345049 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwzqr\" (UniqueName: \"kubernetes.io/projected/ca758fe2-ced2-4ead-95d7-6302c916d2cd-kube-api-access-zwzqr\") pod \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.345184 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-dns-svc\") pod \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.345240 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-config\") pod \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.345305 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-sb\") pod \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\" (UID: \"ca758fe2-ced2-4ead-95d7-6302c916d2cd\") " Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.355596 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca758fe2-ced2-4ead-95d7-6302c916d2cd-kube-api-access-zwzqr" (OuterVolumeSpecName: "kube-api-access-zwzqr") pod "ca758fe2-ced2-4ead-95d7-6302c916d2cd" (UID: "ca758fe2-ced2-4ead-95d7-6302c916d2cd"). InnerVolumeSpecName "kube-api-access-zwzqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.422283 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ca758fe2-ced2-4ead-95d7-6302c916d2cd" (UID: "ca758fe2-ced2-4ead-95d7-6302c916d2cd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.430156 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-config" (OuterVolumeSpecName: "config") pod "ca758fe2-ced2-4ead-95d7-6302c916d2cd" (UID: "ca758fe2-ced2-4ead-95d7-6302c916d2cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.431244 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ca758fe2-ced2-4ead-95d7-6302c916d2cd" (UID: "ca758fe2-ced2-4ead-95d7-6302c916d2cd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.442180 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ca758fe2-ced2-4ead-95d7-6302c916d2cd" (UID: "ca758fe2-ced2-4ead-95d7-6302c916d2cd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.448029 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.448080 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.448095 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.448109 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca758fe2-ced2-4ead-95d7-6302c916d2cd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:13 crc kubenswrapper[4799]: I1124 08:12:13.448124 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwzqr\" (UniqueName: \"kubernetes.io/projected/ca758fe2-ced2-4ead-95d7-6302c916d2cd-kube-api-access-zwzqr\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.117926 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" event={"ID":"ca758fe2-ced2-4ead-95d7-6302c916d2cd","Type":"ContainerDied","Data":"f62d28591840a62aa0adcc217de17894b6230b042d71c6f239c458974e63df69"} Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.118283 4799 scope.go:117] "RemoveContainer" containerID="cf666b3bb466b2cae0b36d18584fb174550d52af5f00880f3ff1c563665e6c6d" Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.118213 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c55bdf6f-249sr" Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.141046 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55c55bdf6f-249sr"] Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.148295 4799 scope.go:117] "RemoveContainer" containerID="b5f774fbaad77232430421a94fc165b44b6e1d361dd436c8b3febb5d75a0826c" Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.150190 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55c55bdf6f-249sr"] Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.261546 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.261754 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="f9749cf2-9a71-453f-88f3-539c15c87c9a" containerName="nova-cell0-conductor-conductor" containerID="cri-o://bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f" gracePeriod=30 Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.284707 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.284927 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2556dfcf-18ac-4b8d-8e92-60784a634311" containerName="nova-scheduler-scheduler" containerID="cri-o://8d95a751a9235a693c1cccdb5e380c98d07e547c691e7e4e81213300871c93a3" gracePeriod=30 Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.293910 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.294209 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerName="nova-api-log" containerID="cri-o://9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad" gracePeriod=30 Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.294370 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerName="nova-api-api" containerID="cri-o://12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958" gracePeriod=30 Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.322070 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.322752 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="8e9c42d8-a576-460f-8629-2536bd6efeba" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://6f1ec4e6c4d0d6987d0877ab3b001517e6bd2387e06ef3e942cffe61e5fb34bc" gracePeriod=30 Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.417132 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.419084 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-log" containerID="cri-o://dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421" gracePeriod=30 Nov 24 08:12:14 crc kubenswrapper[4799]: I1124 08:12:14.419737 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-metadata" containerID="cri-o://e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0" gracePeriod=30 Nov 24 08:12:14 crc kubenswrapper[4799]: E1124 08:12:14.678977 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa72b1d4_0b44_4e83_8f3c_aeec9ac2d192.slice/crio-conmon-dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa72b1d4_0b44_4e83_8f3c_aeec9ac2d192.slice/crio-dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.131165 4799 generic.go:334] "Generic (PLEG): container finished" podID="8e9c42d8-a576-460f-8629-2536bd6efeba" containerID="6f1ec4e6c4d0d6987d0877ab3b001517e6bd2387e06ef3e942cffe61e5fb34bc" exitCode=0 Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.131255 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e9c42d8-a576-460f-8629-2536bd6efeba","Type":"ContainerDied","Data":"6f1ec4e6c4d0d6987d0877ab3b001517e6bd2387e06ef3e942cffe61e5fb34bc"} Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.134614 4799 generic.go:334] "Generic (PLEG): container finished" podID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerID="dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421" exitCode=143 Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.134672 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192","Type":"ContainerDied","Data":"dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421"} Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.140609 4799 generic.go:334] "Generic (PLEG): container finished" podID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerID="9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad" exitCode=143 Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.140653 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f6e1cbd-b387-43fd-9c84-523eb571d041","Type":"ContainerDied","Data":"9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad"} Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.149526 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.265856 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.398420 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-config-data\") pod \"8e9c42d8-a576-460f-8629-2536bd6efeba\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.398509 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-combined-ca-bundle\") pod \"8e9c42d8-a576-460f-8629-2536bd6efeba\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.398726 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m5m2\" (UniqueName: \"kubernetes.io/projected/8e9c42d8-a576-460f-8629-2536bd6efeba-kube-api-access-9m5m2\") pod \"8e9c42d8-a576-460f-8629-2536bd6efeba\" (UID: \"8e9c42d8-a576-460f-8629-2536bd6efeba\") " Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.409514 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e9c42d8-a576-460f-8629-2536bd6efeba-kube-api-access-9m5m2" (OuterVolumeSpecName: "kube-api-access-9m5m2") pod "8e9c42d8-a576-460f-8629-2536bd6efeba" (UID: "8e9c42d8-a576-460f-8629-2536bd6efeba"). InnerVolumeSpecName "kube-api-access-9m5m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.424282 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-config-data" (OuterVolumeSpecName: "config-data") pod "8e9c42d8-a576-460f-8629-2536bd6efeba" (UID: "8e9c42d8-a576-460f-8629-2536bd6efeba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.429464 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e9c42d8-a576-460f-8629-2536bd6efeba" (UID: "8e9c42d8-a576-460f-8629-2536bd6efeba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.500272 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m5m2\" (UniqueName: \"kubernetes.io/projected/8e9c42d8-a576-460f-8629-2536bd6efeba-kube-api-access-9m5m2\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.500305 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.500315 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9c42d8-a576-460f-8629-2536bd6efeba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:15 crc kubenswrapper[4799]: I1124 08:12:15.658646 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca758fe2-ced2-4ead-95d7-6302c916d2cd" path="/var/lib/kubelet/pods/ca758fe2-ced2-4ead-95d7-6302c916d2cd/volumes" Nov 24 08:12:16 crc kubenswrapper[4799]: E1124 08:12:16.138223 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 08:12:16 crc kubenswrapper[4799]: E1124 08:12:16.140521 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 08:12:16 crc kubenswrapper[4799]: E1124 08:12:16.142085 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 08:12:16 crc kubenswrapper[4799]: E1124 08:12:16.142142 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="f9749cf2-9a71-453f-88f3-539c15c87c9a" containerName="nova-cell0-conductor-conductor" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.155606 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e9c42d8-a576-460f-8629-2536bd6efeba","Type":"ContainerDied","Data":"bedea8514d232c85fb8dafa8b9fb80fcb4e62042818182c68cf0144bf4fbadce"} Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.155675 4799 scope.go:117] "RemoveContainer" containerID="6f1ec4e6c4d0d6987d0877ab3b001517e6bd2387e06ef3e942cffe61e5fb34bc" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.155714 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.184189 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.208989 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.223318 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:12:16 crc kubenswrapper[4799]: E1124 08:12:16.223874 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e9c42d8-a576-460f-8629-2536bd6efeba" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.223892 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e9c42d8-a576-460f-8629-2536bd6efeba" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 08:12:16 crc kubenswrapper[4799]: E1124 08:12:16.223950 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca758fe2-ced2-4ead-95d7-6302c916d2cd" containerName="dnsmasq-dns" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.223958 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca758fe2-ced2-4ead-95d7-6302c916d2cd" containerName="dnsmasq-dns" Nov 24 08:12:16 crc kubenswrapper[4799]: E1124 08:12:16.223971 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca758fe2-ced2-4ead-95d7-6302c916d2cd" containerName="init" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.223977 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca758fe2-ced2-4ead-95d7-6302c916d2cd" containerName="init" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.224252 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca758fe2-ced2-4ead-95d7-6302c916d2cd" containerName="dnsmasq-dns" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.224270 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e9c42d8-a576-460f-8629-2536bd6efeba" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.225367 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.229026 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.234196 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.322419 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlfq8\" (UniqueName: \"kubernetes.io/projected/fc1ba947-a1ac-46e3-91a0-f47ead0e6add-kube-api-access-jlfq8\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc1ba947-a1ac-46e3-91a0-f47ead0e6add\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.322593 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc1ba947-a1ac-46e3-91a0-f47ead0e6add-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc1ba947-a1ac-46e3-91a0-f47ead0e6add\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.322654 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc1ba947-a1ac-46e3-91a0-f47ead0e6add-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc1ba947-a1ac-46e3-91a0-f47ead0e6add\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.424116 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlfq8\" (UniqueName: \"kubernetes.io/projected/fc1ba947-a1ac-46e3-91a0-f47ead0e6add-kube-api-access-jlfq8\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc1ba947-a1ac-46e3-91a0-f47ead0e6add\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.424343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc1ba947-a1ac-46e3-91a0-f47ead0e6add-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc1ba947-a1ac-46e3-91a0-f47ead0e6add\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.424391 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc1ba947-a1ac-46e3-91a0-f47ead0e6add-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc1ba947-a1ac-46e3-91a0-f47ead0e6add\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.432048 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc1ba947-a1ac-46e3-91a0-f47ead0e6add-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc1ba947-a1ac-46e3-91a0-f47ead0e6add\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.433007 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc1ba947-a1ac-46e3-91a0-f47ead0e6add-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc1ba947-a1ac-46e3-91a0-f47ead0e6add\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.464888 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlfq8\" (UniqueName: \"kubernetes.io/projected/fc1ba947-a1ac-46e3-91a0-f47ead0e6add-kube-api-access-jlfq8\") pod \"nova-cell1-novncproxy-0\" (UID: \"fc1ba947-a1ac-46e3-91a0-f47ead0e6add\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:16 crc kubenswrapper[4799]: I1124 08:12:16.556247 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:17 crc kubenswrapper[4799]: I1124 08:12:17.046434 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:12:17 crc kubenswrapper[4799]: I1124 08:12:17.167502 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fc1ba947-a1ac-46e3-91a0-f47ead0e6add","Type":"ContainerStarted","Data":"0e80b7fd7b98ee9de725b344ac1c8256b30032d1492a6fed976a3f47d7b8c1d2"} Nov 24 08:12:17 crc kubenswrapper[4799]: E1124 08:12:17.486728 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8d95a751a9235a693c1cccdb5e380c98d07e547c691e7e4e81213300871c93a3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 08:12:17 crc kubenswrapper[4799]: E1124 08:12:17.488670 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8d95a751a9235a693c1cccdb5e380c98d07e547c691e7e4e81213300871c93a3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 08:12:17 crc kubenswrapper[4799]: E1124 08:12:17.490345 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8d95a751a9235a693c1cccdb5e380c98d07e547c691e7e4e81213300871c93a3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 08:12:17 crc kubenswrapper[4799]: E1124 08:12:17.490379 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="2556dfcf-18ac-4b8d-8e92-60784a634311" containerName="nova-scheduler-scheduler" Nov 24 08:12:17 crc kubenswrapper[4799]: I1124 08:12:17.577338 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.71:8775/\": read tcp 10.217.0.2:60084->10.217.1.71:8775: read: connection reset by peer" Nov 24 08:12:17 crc kubenswrapper[4799]: I1124 08:12:17.577396 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.71:8775/\": read tcp 10.217.0.2:60082->10.217.1.71:8775: read: connection reset by peer" Nov 24 08:12:17 crc kubenswrapper[4799]: I1124 08:12:17.639911 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e9c42d8-a576-460f-8629-2536bd6efeba" path="/var/lib/kubelet/pods/8e9c42d8-a576-460f-8629-2536bd6efeba/volumes" Nov 24 08:12:17 crc kubenswrapper[4799]: I1124 08:12:17.784982 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:12:17 crc kubenswrapper[4799]: I1124 08:12:17.785226 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7" containerName="nova-cell1-conductor-conductor" containerID="cri-o://8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511" gracePeriod=30 Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.034976 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.087895 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.175264 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgsdk\" (UniqueName: \"kubernetes.io/projected/3f6e1cbd-b387-43fd-9c84-523eb571d041-kube-api-access-pgsdk\") pod \"3f6e1cbd-b387-43fd-9c84-523eb571d041\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.175346 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85s5g\" (UniqueName: \"kubernetes.io/projected/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-kube-api-access-85s5g\") pod \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.175414 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-combined-ca-bundle\") pod \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.175477 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-config-data\") pod \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.175534 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f6e1cbd-b387-43fd-9c84-523eb571d041-logs\") pod \"3f6e1cbd-b387-43fd-9c84-523eb571d041\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.175562 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-config-data\") pod \"3f6e1cbd-b387-43fd-9c84-523eb571d041\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.175616 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-combined-ca-bundle\") pod \"3f6e1cbd-b387-43fd-9c84-523eb571d041\" (UID: \"3f6e1cbd-b387-43fd-9c84-523eb571d041\") " Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.175671 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-logs\") pod \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\" (UID: \"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192\") " Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.176483 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-logs" (OuterVolumeSpecName: "logs") pod "fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" (UID: "fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.177301 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f6e1cbd-b387-43fd-9c84-523eb571d041-logs" (OuterVolumeSpecName: "logs") pod "3f6e1cbd-b387-43fd-9c84-523eb571d041" (UID: "3f6e1cbd-b387-43fd-9c84-523eb571d041"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.182143 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-kube-api-access-85s5g" (OuterVolumeSpecName: "kube-api-access-85s5g") pod "fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" (UID: "fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192"). InnerVolumeSpecName "kube-api-access-85s5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.183359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fc1ba947-a1ac-46e3-91a0-f47ead0e6add","Type":"ContainerStarted","Data":"6e6d85f8e80310eee8d49a3397cb0c5b9e11cec70992be6a281b2363727a299b"} Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.190068 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f6e1cbd-b387-43fd-9c84-523eb571d041-kube-api-access-pgsdk" (OuterVolumeSpecName: "kube-api-access-pgsdk") pod "3f6e1cbd-b387-43fd-9c84-523eb571d041" (UID: "3f6e1cbd-b387-43fd-9c84-523eb571d041"). InnerVolumeSpecName "kube-api-access-pgsdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.202088 4799 generic.go:334] "Generic (PLEG): container finished" podID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerID="12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958" exitCode=0 Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.202184 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f6e1cbd-b387-43fd-9c84-523eb571d041","Type":"ContainerDied","Data":"12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958"} Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.202213 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f6e1cbd-b387-43fd-9c84-523eb571d041","Type":"ContainerDied","Data":"514f6995a94f27b8cba93b8ed2d48ebc3d392cade7a5b3363166147bdce1f6e0"} Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.202259 4799 scope.go:117] "RemoveContainer" containerID="12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.202578 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.202890 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f6e1cbd-b387-43fd-9c84-523eb571d041" (UID: "3f6e1cbd-b387-43fd-9c84-523eb571d041"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.211127 4799 generic.go:334] "Generic (PLEG): container finished" podID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerID="e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0" exitCode=0 Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.211167 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192","Type":"ContainerDied","Data":"e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0"} Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.211195 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192","Type":"ContainerDied","Data":"f3031ac51a9baae4f7d516502477ba3bf0f209003b6a89f01c53819fc8935472"} Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.211253 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.216770 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.216750699 podStartE2EDuration="2.216750699s" podCreationTimestamp="2025-11-24 08:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:12:18.205801399 +0000 UTC m=+5083.861783903" watchObservedRunningTime="2025-11-24 08:12:18.216750699 +0000 UTC m=+5083.872733163" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.228273 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-config-data" (OuterVolumeSpecName: "config-data") pod "3f6e1cbd-b387-43fd-9c84-523eb571d041" (UID: "3f6e1cbd-b387-43fd-9c84-523eb571d041"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.234791 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-config-data" (OuterVolumeSpecName: "config-data") pod "fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" (UID: "fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.234968 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" (UID: "fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.248679 4799 scope.go:117] "RemoveContainer" containerID="9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.279369 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85s5g\" (UniqueName: \"kubernetes.io/projected/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-kube-api-access-85s5g\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.279406 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.279420 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.279433 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f6e1cbd-b387-43fd-9c84-523eb571d041-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.279447 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.279459 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6e1cbd-b387-43fd-9c84-523eb571d041-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.279470 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.279483 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgsdk\" (UniqueName: \"kubernetes.io/projected/3f6e1cbd-b387-43fd-9c84-523eb571d041-kube-api-access-pgsdk\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.295333 4799 scope.go:117] "RemoveContainer" containerID="12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958" Nov 24 08:12:18 crc kubenswrapper[4799]: E1124 08:12:18.295774 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958\": container with ID starting with 12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958 not found: ID does not exist" containerID="12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.295814 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958"} err="failed to get container status \"12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958\": rpc error: code = NotFound desc = could not find container \"12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958\": container with ID starting with 12a40fbcf24aae007ac6c216d748835f9b264147b1c843c865d2065bec73c958 not found: ID does not exist" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.295832 4799 scope.go:117] "RemoveContainer" containerID="9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad" Nov 24 08:12:18 crc kubenswrapper[4799]: E1124 08:12:18.296222 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad\": container with ID starting with 9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad not found: ID does not exist" containerID="9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.296316 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad"} err="failed to get container status \"9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad\": rpc error: code = NotFound desc = could not find container \"9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad\": container with ID starting with 9ab5703a8d145c83ca4a1e02b4f4c660dc745df1c2ab412aff3f3bf083cd1cad not found: ID does not exist" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.296352 4799 scope.go:117] "RemoveContainer" containerID="e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.323921 4799 scope.go:117] "RemoveContainer" containerID="dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.347242 4799 scope.go:117] "RemoveContainer" containerID="e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0" Nov 24 08:12:18 crc kubenswrapper[4799]: E1124 08:12:18.347770 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0\": container with ID starting with e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0 not found: ID does not exist" containerID="e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.347817 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0"} err="failed to get container status \"e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0\": rpc error: code = NotFound desc = could not find container \"e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0\": container with ID starting with e2b48964caa7b66e67fc5b1fc60e7fef979724a239e5eb3507a274c0db6c73c0 not found: ID does not exist" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.347867 4799 scope.go:117] "RemoveContainer" containerID="dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421" Nov 24 08:12:18 crc kubenswrapper[4799]: E1124 08:12:18.348398 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421\": container with ID starting with dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421 not found: ID does not exist" containerID="dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.348455 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421"} err="failed to get container status \"dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421\": rpc error: code = NotFound desc = could not find container \"dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421\": container with ID starting with dfc8d71327223cef46b90530ba834c93a21837c24319ce6dcacd0aa9fc37f421 not found: ID does not exist" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.549258 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.569476 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.581675 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.588823 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.599490 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 08:12:18 crc kubenswrapper[4799]: E1124 08:12:18.599902 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerName="nova-api-log" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.599922 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerName="nova-api-log" Nov 24 08:12:18 crc kubenswrapper[4799]: E1124 08:12:18.599939 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-metadata" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.599959 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-metadata" Nov 24 08:12:18 crc kubenswrapper[4799]: E1124 08:12:18.599974 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-log" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.599980 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-log" Nov 24 08:12:18 crc kubenswrapper[4799]: E1124 08:12:18.599995 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerName="nova-api-api" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.600001 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerName="nova-api-api" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.600199 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerName="nova-api-api" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.600223 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-metadata" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.600233 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f6e1cbd-b387-43fd-9c84-523eb571d041" containerName="nova-api-log" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.600243 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" containerName="nova-metadata-log" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.601330 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.603570 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.613685 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.623407 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.627257 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.633378 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.637364 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.688116 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-config-data\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.688165 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27fb95fd-a5be-4c99-9ce3-f8617695a530-logs\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.688212 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lms8v\" (UniqueName: \"kubernetes.io/projected/3966b7b5-c531-4127-82d0-8d759464708a-kube-api-access-lms8v\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.688239 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvjcp\" (UniqueName: \"kubernetes.io/projected/27fb95fd-a5be-4c99-9ce3-f8617695a530-kube-api-access-nvjcp\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.688260 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.688296 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-config-data\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.688314 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.688338 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3966b7b5-c531-4127-82d0-8d759464708a-logs\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.789771 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-config-data\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.789822 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.789865 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3966b7b5-c531-4127-82d0-8d759464708a-logs\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.789936 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-config-data\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.789956 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27fb95fd-a5be-4c99-9ce3-f8617695a530-logs\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.789998 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lms8v\" (UniqueName: \"kubernetes.io/projected/3966b7b5-c531-4127-82d0-8d759464708a-kube-api-access-lms8v\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.790029 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvjcp\" (UniqueName: \"kubernetes.io/projected/27fb95fd-a5be-4c99-9ce3-f8617695a530-kube-api-access-nvjcp\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.790051 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.790527 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27fb95fd-a5be-4c99-9ce3-f8617695a530-logs\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.790593 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3966b7b5-c531-4127-82d0-8d759464708a-logs\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.793919 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.794062 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.796200 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-config-data\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.801500 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-config-data\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.821328 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lms8v\" (UniqueName: \"kubernetes.io/projected/3966b7b5-c531-4127-82d0-8d759464708a-kube-api-access-lms8v\") pod \"nova-api-0\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.821919 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvjcp\" (UniqueName: \"kubernetes.io/projected/27fb95fd-a5be-4c99-9ce3-f8617695a530-kube-api-access-nvjcp\") pod \"nova-metadata-0\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " pod="openstack/nova-metadata-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.918684 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:12:18 crc kubenswrapper[4799]: I1124 08:12:18.961211 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:12:19 crc kubenswrapper[4799]: I1124 08:12:19.406809 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:12:19 crc kubenswrapper[4799]: W1124 08:12:19.408873 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3966b7b5_c531_4127_82d0_8d759464708a.slice/crio-59c3bcdd109c1136c39115f7eb46b0a21614e3f1e2e2946d8e92b0a48d2b8a65 WatchSource:0}: Error finding container 59c3bcdd109c1136c39115f7eb46b0a21614e3f1e2e2946d8e92b0a48d2b8a65: Status 404 returned error can't find the container with id 59c3bcdd109c1136c39115f7eb46b0a21614e3f1e2e2946d8e92b0a48d2b8a65 Nov 24 08:12:19 crc kubenswrapper[4799]: I1124 08:12:19.483616 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:12:19 crc kubenswrapper[4799]: I1124 08:12:19.642423 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f6e1cbd-b387-43fd-9c84-523eb571d041" path="/var/lib/kubelet/pods/3f6e1cbd-b387-43fd-9c84-523eb571d041/volumes" Nov 24 08:12:19 crc kubenswrapper[4799]: I1124 08:12:19.643620 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192" path="/var/lib/kubelet/pods/fa72b1d4-0b44-4e83-8f3c-aeec9ac2d192/volumes" Nov 24 08:12:19 crc kubenswrapper[4799]: I1124 08:12:19.973531 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.012653 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-config-data\") pod \"f9749cf2-9a71-453f-88f3-539c15c87c9a\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.012743 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-combined-ca-bundle\") pod \"f9749cf2-9a71-453f-88f3-539c15c87c9a\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.012828 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnxv6\" (UniqueName: \"kubernetes.io/projected/f9749cf2-9a71-453f-88f3-539c15c87c9a-kube-api-access-wnxv6\") pod \"f9749cf2-9a71-453f-88f3-539c15c87c9a\" (UID: \"f9749cf2-9a71-453f-88f3-539c15c87c9a\") " Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.017010 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9749cf2-9a71-453f-88f3-539c15c87c9a-kube-api-access-wnxv6" (OuterVolumeSpecName: "kube-api-access-wnxv6") pod "f9749cf2-9a71-453f-88f3-539c15c87c9a" (UID: "f9749cf2-9a71-453f-88f3-539c15c87c9a"). InnerVolumeSpecName "kube-api-access-wnxv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.058137 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-config-data" (OuterVolumeSpecName: "config-data") pod "f9749cf2-9a71-453f-88f3-539c15c87c9a" (UID: "f9749cf2-9a71-453f-88f3-539c15c87c9a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.069190 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9749cf2-9a71-453f-88f3-539c15c87c9a" (UID: "f9749cf2-9a71-453f-88f3-539c15c87c9a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.115156 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.115503 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9749cf2-9a71-453f-88f3-539c15c87c9a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.115515 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnxv6\" (UniqueName: \"kubernetes.io/projected/f9749cf2-9a71-453f-88f3-539c15c87c9a-kube-api-access-wnxv6\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.235343 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27fb95fd-a5be-4c99-9ce3-f8617695a530","Type":"ContainerStarted","Data":"bae71ae1cf9e4b2ec05692bd4e70b2a8132ec02f03bac45d32f8a3fe145198a9"} Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.235385 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27fb95fd-a5be-4c99-9ce3-f8617695a530","Type":"ContainerStarted","Data":"de6e817697e87c1221c2c1aa6845284ecd7b05c5c743fa80dc4ce0ed79f8fbc0"} Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.235395 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27fb95fd-a5be-4c99-9ce3-f8617695a530","Type":"ContainerStarted","Data":"f19e33828d5e640ad17a5d1eaa9c2924d4a5d2d0adad9ffae4eeebd7b0448b17"} Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.238264 4799 generic.go:334] "Generic (PLEG): container finished" podID="f9749cf2-9a71-453f-88f3-539c15c87c9a" containerID="bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f" exitCode=0 Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.238316 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f9749cf2-9a71-453f-88f3-539c15c87c9a","Type":"ContainerDied","Data":"bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f"} Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.238334 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f9749cf2-9a71-453f-88f3-539c15c87c9a","Type":"ContainerDied","Data":"7051e3f81fa53419831335302245ec1579cf677d3f1d71c63875bac4472033ff"} Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.238350 4799 scope.go:117] "RemoveContainer" containerID="bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.238447 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.249576 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3966b7b5-c531-4127-82d0-8d759464708a","Type":"ContainerStarted","Data":"53b35860f548f35f4338efbdd33b9ee83820c2b24af2065c9e20cbe4f405183b"} Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.249629 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3966b7b5-c531-4127-82d0-8d759464708a","Type":"ContainerStarted","Data":"46251192c28f8377b93fd73ac7b5fda79e3f27f7ede96780c4906f47d9f1d9e3"} Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.249643 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3966b7b5-c531-4127-82d0-8d759464708a","Type":"ContainerStarted","Data":"59c3bcdd109c1136c39115f7eb46b0a21614e3f1e2e2946d8e92b0a48d2b8a65"} Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.270700 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.270681806 podStartE2EDuration="2.270681806s" podCreationTimestamp="2025-11-24 08:12:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:12:20.264274404 +0000 UTC m=+5085.920256878" watchObservedRunningTime="2025-11-24 08:12:20.270681806 +0000 UTC m=+5085.926664280" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.295643 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.295620364 podStartE2EDuration="2.295620364s" podCreationTimestamp="2025-11-24 08:12:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:12:20.283810009 +0000 UTC m=+5085.939792493" watchObservedRunningTime="2025-11-24 08:12:20.295620364 +0000 UTC m=+5085.951602838" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.301090 4799 scope.go:117] "RemoveContainer" containerID="bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f" Nov 24 08:12:20 crc kubenswrapper[4799]: E1124 08:12:20.301549 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f\": container with ID starting with bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f not found: ID does not exist" containerID="bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.301592 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f"} err="failed to get container status \"bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f\": rpc error: code = NotFound desc = could not find container \"bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f\": container with ID starting with bff693408576ab9dcbd82c9340a6d4e84822f843769aad00b7548141f74b4d9f not found: ID does not exist" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.311645 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.330021 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.341685 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:12:20 crc kubenswrapper[4799]: E1124 08:12:20.342130 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9749cf2-9a71-453f-88f3-539c15c87c9a" containerName="nova-cell0-conductor-conductor" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.342147 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9749cf2-9a71-453f-88f3-539c15c87c9a" containerName="nova-cell0-conductor-conductor" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.342323 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9749cf2-9a71-453f-88f3-539c15c87c9a" containerName="nova-cell0-conductor-conductor" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.342954 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.350094 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.354059 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.420989 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.421097 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssvlh\" (UniqueName: \"kubernetes.io/projected/2516f20b-d128-4383-9982-699005340d35-kube-api-access-ssvlh\") pod \"nova-cell0-conductor-0\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.421177 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.523424 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.523517 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssvlh\" (UniqueName: \"kubernetes.io/projected/2516f20b-d128-4383-9982-699005340d35-kube-api-access-ssvlh\") pod \"nova-cell0-conductor-0\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.523585 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.528331 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.530204 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.542617 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssvlh\" (UniqueName: \"kubernetes.io/projected/2516f20b-d128-4383-9982-699005340d35-kube-api-access-ssvlh\") pod \"nova-cell0-conductor-0\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.670554 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.892091 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.935404 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-combined-ca-bundle\") pod \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.935476 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-config-data\") pod \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.935532 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdkq5\" (UniqueName: \"kubernetes.io/projected/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-kube-api-access-hdkq5\") pod \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\" (UID: \"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7\") " Nov 24 08:12:20 crc kubenswrapper[4799]: I1124 08:12:20.971518 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-kube-api-access-hdkq5" (OuterVolumeSpecName: "kube-api-access-hdkq5") pod "8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7" (UID: "8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7"). InnerVolumeSpecName "kube-api-access-hdkq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.020741 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-config-data" (OuterVolumeSpecName: "config-data") pod "8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7" (UID: "8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.036638 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7" (UID: "8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.038181 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.038201 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.038209 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdkq5\" (UniqueName: \"kubernetes.io/projected/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7-kube-api-access-hdkq5\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.263968 4799 generic.go:334] "Generic (PLEG): container finished" podID="2556dfcf-18ac-4b8d-8e92-60784a634311" containerID="8d95a751a9235a693c1cccdb5e380c98d07e547c691e7e4e81213300871c93a3" exitCode=0 Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.264276 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2556dfcf-18ac-4b8d-8e92-60784a634311","Type":"ContainerDied","Data":"8d95a751a9235a693c1cccdb5e380c98d07e547c691e7e4e81213300871c93a3"} Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.269802 4799 generic.go:334] "Generic (PLEG): container finished" podID="8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7" containerID="8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511" exitCode=0 Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.269885 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.269927 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7","Type":"ContainerDied","Data":"8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511"} Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.269959 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7","Type":"ContainerDied","Data":"7899e49b55df1c5e848aab827269fac06fb78ab49880d1621fb4b6dfc1c3cbce"} Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.269979 4799 scope.go:117] "RemoveContainer" containerID="8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.335086 4799 scope.go:117] "RemoveContainer" containerID="8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.335249 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:12:21 crc kubenswrapper[4799]: E1124 08:12:21.336451 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511\": container with ID starting with 8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511 not found: ID does not exist" containerID="8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.336496 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511"} err="failed to get container status \"8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511\": rpc error: code = NotFound desc = could not find container \"8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511\": container with ID starting with 8037a215342fe1d2df699d74773f7ca213b344921261e4718f539f6c9dfe0511 not found: ID does not exist" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.356775 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.373420 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:12:21 crc kubenswrapper[4799]: E1124 08:12:21.374008 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7" containerName="nova-cell1-conductor-conductor" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.374033 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7" containerName="nova-cell1-conductor-conductor" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.374248 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7" containerName="nova-cell1-conductor-conductor" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.375487 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.377930 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.395817 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:12:21 crc kubenswrapper[4799]: W1124 08:12:21.412450 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2516f20b_d128_4383_9982_699005340d35.slice/crio-839e1daf35d8c2935d8a5f1329d56edab47fe9130467f0a56867b910bc467355 WatchSource:0}: Error finding container 839e1daf35d8c2935d8a5f1329d56edab47fe9130467f0a56867b910bc467355: Status 404 returned error can't find the container with id 839e1daf35d8c2935d8a5f1329d56edab47fe9130467f0a56867b910bc467355 Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.414587 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.430763 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.451150 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ffvr\" (UniqueName: \"kubernetes.io/projected/2556dfcf-18ac-4b8d-8e92-60784a634311-kube-api-access-8ffvr\") pod \"2556dfcf-18ac-4b8d-8e92-60784a634311\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.451320 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-combined-ca-bundle\") pod \"2556dfcf-18ac-4b8d-8e92-60784a634311\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.451443 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-config-data\") pod \"2556dfcf-18ac-4b8d-8e92-60784a634311\" (UID: \"2556dfcf-18ac-4b8d-8e92-60784a634311\") " Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.451656 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.451693 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.451873 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkghd\" (UniqueName: \"kubernetes.io/projected/2ceaea61-e8ed-455e-8020-b96783d44e72-kube-api-access-vkghd\") pod \"nova-cell1-conductor-0\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.455047 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2556dfcf-18ac-4b8d-8e92-60784a634311-kube-api-access-8ffvr" (OuterVolumeSpecName: "kube-api-access-8ffvr") pod "2556dfcf-18ac-4b8d-8e92-60784a634311" (UID: "2556dfcf-18ac-4b8d-8e92-60784a634311"). InnerVolumeSpecName "kube-api-access-8ffvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.482995 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2556dfcf-18ac-4b8d-8e92-60784a634311" (UID: "2556dfcf-18ac-4b8d-8e92-60784a634311"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.493587 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-config-data" (OuterVolumeSpecName: "config-data") pod "2556dfcf-18ac-4b8d-8e92-60784a634311" (UID: "2556dfcf-18ac-4b8d-8e92-60784a634311"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.553799 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkghd\" (UniqueName: \"kubernetes.io/projected/2ceaea61-e8ed-455e-8020-b96783d44e72-kube-api-access-vkghd\") pod \"nova-cell1-conductor-0\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.553961 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.554000 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.554096 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.554119 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ffvr\" (UniqueName: \"kubernetes.io/projected/2556dfcf-18ac-4b8d-8e92-60784a634311-kube-api-access-8ffvr\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.554132 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2556dfcf-18ac-4b8d-8e92-60784a634311-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.557049 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.557797 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.559006 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.574243 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkghd\" (UniqueName: \"kubernetes.io/projected/2ceaea61-e8ed-455e-8020-b96783d44e72-kube-api-access-vkghd\") pod \"nova-cell1-conductor-0\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.642027 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7" path="/var/lib/kubelet/pods/8d01fe98-c3d6-45b2-83a8-4ef1f00ec3e7/volumes" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.643413 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9749cf2-9a71-453f-88f3-539c15c87c9a" path="/var/lib/kubelet/pods/f9749cf2-9a71-453f-88f3-539c15c87c9a/volumes" Nov 24 08:12:21 crc kubenswrapper[4799]: I1124 08:12:21.700081 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.153458 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.286748 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2ceaea61-e8ed-455e-8020-b96783d44e72","Type":"ContainerStarted","Data":"6e512e3e417266c72dc43025aeb8293a5530347cbce2891afb45fd39a2a266bf"} Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.288708 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2556dfcf-18ac-4b8d-8e92-60784a634311","Type":"ContainerDied","Data":"f2c25f4afececd9e4afa9df6dd86dba448b639100e9f8a6711286413f5d40004"} Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.288827 4799 scope.go:117] "RemoveContainer" containerID="8d95a751a9235a693c1cccdb5e380c98d07e547c691e7e4e81213300871c93a3" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.289110 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.292660 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2516f20b-d128-4383-9982-699005340d35","Type":"ContainerStarted","Data":"75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286"} Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.292733 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2516f20b-d128-4383-9982-699005340d35","Type":"ContainerStarted","Data":"839e1daf35d8c2935d8a5f1329d56edab47fe9130467f0a56867b910bc467355"} Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.292797 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.317378 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.317348938 podStartE2EDuration="2.317348938s" podCreationTimestamp="2025-11-24 08:12:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:12:22.313111378 +0000 UTC m=+5087.969093872" watchObservedRunningTime="2025-11-24 08:12:22.317348938 +0000 UTC m=+5087.973331432" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.339385 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.347329 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.378868 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:12:22 crc kubenswrapper[4799]: E1124 08:12:22.379200 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2556dfcf-18ac-4b8d-8e92-60784a634311" containerName="nova-scheduler-scheduler" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.379216 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2556dfcf-18ac-4b8d-8e92-60784a634311" containerName="nova-scheduler-scheduler" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.379390 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2556dfcf-18ac-4b8d-8e92-60784a634311" containerName="nova-scheduler-scheduler" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.379917 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.379983 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.382563 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.480436 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhgzs\" (UniqueName: \"kubernetes.io/projected/19134f4a-118d-4e1c-b53d-742d563576c0-kube-api-access-dhgzs\") pod \"nova-scheduler-0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " pod="openstack/nova-scheduler-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.481611 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-config-data\") pod \"nova-scheduler-0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " pod="openstack/nova-scheduler-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.481800 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " pod="openstack/nova-scheduler-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.583569 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " pod="openstack/nova-scheduler-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.584119 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhgzs\" (UniqueName: \"kubernetes.io/projected/19134f4a-118d-4e1c-b53d-742d563576c0-kube-api-access-dhgzs\") pod \"nova-scheduler-0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " pod="openstack/nova-scheduler-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.584286 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-config-data\") pod \"nova-scheduler-0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " pod="openstack/nova-scheduler-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.588746 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-config-data\") pod \"nova-scheduler-0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " pod="openstack/nova-scheduler-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.589342 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " pod="openstack/nova-scheduler-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.600323 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhgzs\" (UniqueName: \"kubernetes.io/projected/19134f4a-118d-4e1c-b53d-742d563576c0-kube-api-access-dhgzs\") pod \"nova-scheduler-0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " pod="openstack/nova-scheduler-0" Nov 24 08:12:22 crc kubenswrapper[4799]: I1124 08:12:22.708033 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:12:23 crc kubenswrapper[4799]: W1124 08:12:23.220970 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19134f4a_118d_4e1c_b53d_742d563576c0.slice/crio-319e8891665d9dd4f986476ebbf101d669103d15f4b7224aa2d2cf6712aff7db WatchSource:0}: Error finding container 319e8891665d9dd4f986476ebbf101d669103d15f4b7224aa2d2cf6712aff7db: Status 404 returned error can't find the container with id 319e8891665d9dd4f986476ebbf101d669103d15f4b7224aa2d2cf6712aff7db Nov 24 08:12:23 crc kubenswrapper[4799]: I1124 08:12:23.227864 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:12:23 crc kubenswrapper[4799]: I1124 08:12:23.303795 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2ceaea61-e8ed-455e-8020-b96783d44e72","Type":"ContainerStarted","Data":"fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7"} Nov 24 08:12:23 crc kubenswrapper[4799]: I1124 08:12:23.304447 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:23 crc kubenswrapper[4799]: I1124 08:12:23.306316 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19134f4a-118d-4e1c-b53d-742d563576c0","Type":"ContainerStarted","Data":"319e8891665d9dd4f986476ebbf101d669103d15f4b7224aa2d2cf6712aff7db"} Nov 24 08:12:23 crc kubenswrapper[4799]: I1124 08:12:23.327510 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.327490223 podStartE2EDuration="2.327490223s" podCreationTimestamp="2025-11-24 08:12:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:12:23.317712736 +0000 UTC m=+5088.973695220" watchObservedRunningTime="2025-11-24 08:12:23.327490223 +0000 UTC m=+5088.983472697" Nov 24 08:12:23 crc kubenswrapper[4799]: I1124 08:12:23.628837 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:12:23 crc kubenswrapper[4799]: E1124 08:12:23.634721 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:12:23 crc kubenswrapper[4799]: I1124 08:12:23.642905 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2556dfcf-18ac-4b8d-8e92-60784a634311" path="/var/lib/kubelet/pods/2556dfcf-18ac-4b8d-8e92-60784a634311/volumes" Nov 24 08:12:23 crc kubenswrapper[4799]: I1124 08:12:23.961762 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:12:23 crc kubenswrapper[4799]: I1124 08:12:23.962076 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:12:24 crc kubenswrapper[4799]: I1124 08:12:24.319785 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19134f4a-118d-4e1c-b53d-742d563576c0","Type":"ContainerStarted","Data":"4eca8afcb5ad58ce56463cb107cd2c212cce92b084a87e47d7e2c55e7954ea6c"} Nov 24 08:12:24 crc kubenswrapper[4799]: I1124 08:12:24.340670 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.340648695 podStartE2EDuration="2.340648695s" podCreationTimestamp="2025-11-24 08:12:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:12:24.332470733 +0000 UTC m=+5089.988453207" watchObservedRunningTime="2025-11-24 08:12:24.340648695 +0000 UTC m=+5089.996631179" Nov 24 08:12:26 crc kubenswrapper[4799]: I1124 08:12:26.557405 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:26 crc kubenswrapper[4799]: I1124 08:12:26.570553 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:27 crc kubenswrapper[4799]: I1124 08:12:27.398730 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:12:27 crc kubenswrapper[4799]: I1124 08:12:27.711943 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 08:12:28 crc kubenswrapper[4799]: I1124 08:12:28.919135 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:12:28 crc kubenswrapper[4799]: I1124 08:12:28.919596 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:12:28 crc kubenswrapper[4799]: I1124 08:12:28.961618 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 08:12:28 crc kubenswrapper[4799]: I1124 08:12:28.961708 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 08:12:30 crc kubenswrapper[4799]: I1124 08:12:30.001044 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3966b7b5-c531-4127-82d0-8d759464708a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.80:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:12:30 crc kubenswrapper[4799]: I1124 08:12:30.001109 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3966b7b5-c531-4127-82d0-8d759464708a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.80:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:12:30 crc kubenswrapper[4799]: I1124 08:12:30.083159 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:12:30 crc kubenswrapper[4799]: I1124 08:12:30.083456 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:12:30 crc kubenswrapper[4799]: I1124 08:12:30.702557 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 08:12:31 crc kubenswrapper[4799]: I1124 08:12:31.730410 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 08:12:32 crc kubenswrapper[4799]: I1124 08:12:32.712468 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 08:12:32 crc kubenswrapper[4799]: I1124 08:12:32.738206 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.250154 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.252247 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.254271 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.261839 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.303591 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.303702 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.303868 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.303951 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-scripts\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.303980 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vqvk\" (UniqueName: \"kubernetes.io/projected/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-kube-api-access-6vqvk\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.304029 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.406112 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.406204 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-scripts\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.406243 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vqvk\" (UniqueName: \"kubernetes.io/projected/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-kube-api-access-6vqvk\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.406277 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.406277 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.406312 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.406464 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.413463 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-scripts\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.413492 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.413815 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.418449 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.429649 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vqvk\" (UniqueName: \"kubernetes.io/projected/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-kube-api-access-6vqvk\") pod \"cinder-scheduler-0\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.467482 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 08:12:33 crc kubenswrapper[4799]: I1124 08:12:33.594413 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.134995 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.315463 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.316195 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7f05f65a-f0e1-4645-8085-5614e23dc198" containerName="cinder-api-log" containerID="cri-o://db78d862d8e4edd6078acf7c227757b2f6bd51963574eecb371a2dfe75e3b3bc" gracePeriod=30 Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.316264 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7f05f65a-f0e1-4645-8085-5614e23dc198" containerName="cinder-api" containerID="cri-o://2a794cfe903d6ddfe89e23172c37df55f73ab3305a535a8ea7f5d0413d7a414b" gracePeriod=30 Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.447990 4799 generic.go:334] "Generic (PLEG): container finished" podID="7f05f65a-f0e1-4645-8085-5614e23dc198" containerID="db78d862d8e4edd6078acf7c227757b2f6bd51963574eecb371a2dfe75e3b3bc" exitCode=143 Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.448070 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f05f65a-f0e1-4645-8085-5614e23dc198","Type":"ContainerDied","Data":"db78d862d8e4edd6078acf7c227757b2f6bd51963574eecb371a2dfe75e3b3bc"} Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.450222 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569","Type":"ContainerStarted","Data":"046ff57d77d3f9c4530240fe1c12aeaeb91ca4346688447e15d6d5d2cad3ce9a"} Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.629525 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:12:34 crc kubenswrapper[4799]: E1124 08:12:34.629732 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.833031 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.837096 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.840086 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.847317 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.941017 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-sys\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.941281 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-dev\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.941538 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.941656 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kmzp\" (UniqueName: \"kubernetes.io/projected/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-kube-api-access-9kmzp\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.941771 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.941928 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.942054 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.942181 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.942271 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.942353 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.942434 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.942511 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-run\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.942619 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.942713 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.942788 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:34 crc kubenswrapper[4799]: I1124 08:12:34.942888 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.043983 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.044309 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.044388 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.044461 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-run\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.044552 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.044634 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.044702 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.044779 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.044909 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-sys\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.044995 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-dev\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.045103 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.045193 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kmzp\" (UniqueName: \"kubernetes.io/projected/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-kube-api-access-9kmzp\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.045276 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.045361 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.045435 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.045512 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.045812 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.046792 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-run\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.047009 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.047156 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.047502 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-sys\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.047603 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-dev\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.048052 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.048501 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.050156 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.050296 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.052793 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.059517 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.060210 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.061527 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.062223 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.072425 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kmzp\" (UniqueName: \"kubernetes.io/projected/b55c4e6c-8c8b-4623-b420-3cfb9eea5764-kube-api-access-9kmzp\") pod \"cinder-volume-volume1-0\" (UID: \"b55c4e6c-8c8b-4623-b420-3cfb9eea5764\") " pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.216014 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.480119 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.482349 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.484655 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.513043 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.535177 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569","Type":"ContainerStarted","Data":"53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf"} Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.660311 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-dev\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661058 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661096 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661121 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661479 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-lib-modules\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661522 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2c91ef74-c8d3-4283-913d-16cd52f50546-ceph\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661556 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661587 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661633 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-run\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661742 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-config-data\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661840 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-etc-nvme\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661893 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-sys\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.661983 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-scripts\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.662006 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.662037 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-config-data-custom\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.662078 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z985\" (UniqueName: \"kubernetes.io/projected/2c91ef74-c8d3-4283-913d-16cd52f50546-kube-api-access-9z985\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.757588 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.763764 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.764113 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-lib-modules\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.764928 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.763947 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-lib-modules\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765091 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2c91ef74-c8d3-4283-913d-16cd52f50546-ceph\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765130 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765189 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765237 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-run\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765308 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-config-data\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765363 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-etc-nvme\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765385 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-sys\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765429 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-scripts\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765449 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765468 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-config-data-custom\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765499 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z985\" (UniqueName: \"kubernetes.io/projected/2c91ef74-c8d3-4283-913d-16cd52f50546-kube-api-access-9z985\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765520 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-dev\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765536 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.765570 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.766903 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.766908 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-etc-nvme\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.766983 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.767858 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-run\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.767928 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.767986 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-sys\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.768045 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-dev\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.768239 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2c91ef74-c8d3-4283-913d-16cd52f50546-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.771023 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-scripts\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.771656 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.773020 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2c91ef74-c8d3-4283-913d-16cd52f50546-ceph\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.774976 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-config-data-custom\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.782974 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c91ef74-c8d3-4283-913d-16cd52f50546-config-data\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.784737 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z985\" (UniqueName: \"kubernetes.io/projected/2c91ef74-c8d3-4283-913d-16cd52f50546-kube-api-access-9z985\") pod \"cinder-backup-0\" (UID: \"2c91ef74-c8d3-4283-913d-16cd52f50546\") " pod="openstack/cinder-backup-0" Nov 24 08:12:35 crc kubenswrapper[4799]: I1124 08:12:35.814949 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 24 08:12:36 crc kubenswrapper[4799]: I1124 08:12:36.400053 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 24 08:12:36 crc kubenswrapper[4799]: I1124 08:12:36.553107 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"b55c4e6c-8c8b-4623-b420-3cfb9eea5764","Type":"ContainerStarted","Data":"9092ae2645ba6483f98c9e9ca4bbb6fc31e05bed8cbff60ce4cabe13e6839b4d"} Nov 24 08:12:36 crc kubenswrapper[4799]: I1124 08:12:36.553456 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"b55c4e6c-8c8b-4623-b420-3cfb9eea5764","Type":"ContainerStarted","Data":"5d49d212b2f3f9026ea45127a6be9cfc1e7a531913b17b6458d7fa4175c048eb"} Nov 24 08:12:36 crc kubenswrapper[4799]: I1124 08:12:36.555380 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569","Type":"ContainerStarted","Data":"4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad"} Nov 24 08:12:36 crc kubenswrapper[4799]: I1124 08:12:36.570710 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"2c91ef74-c8d3-4283-913d-16cd52f50546","Type":"ContainerStarted","Data":"dbd5da34f499681ba1e371a1af0ead6e70e034a28c2ce7103902cdef614209e8"} Nov 24 08:12:36 crc kubenswrapper[4799]: I1124 08:12:36.580306 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.255397802 podStartE2EDuration="3.580288246s" podCreationTimestamp="2025-11-24 08:12:33 +0000 UTC" firstStartedPulling="2025-11-24 08:12:34.142332066 +0000 UTC m=+5099.798314560" lastFinishedPulling="2025-11-24 08:12:34.46722253 +0000 UTC m=+5100.123205004" observedRunningTime="2025-11-24 08:12:36.576386595 +0000 UTC m=+5102.232369059" watchObservedRunningTime="2025-11-24 08:12:36.580288246 +0000 UTC m=+5102.236270720" Nov 24 08:12:37 crc kubenswrapper[4799]: I1124 08:12:37.588305 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"2c91ef74-c8d3-4283-913d-16cd52f50546","Type":"ContainerStarted","Data":"8480a1bf55026b747e9337fc8f176d38ed56dd19754093afb0e50d55e243dc47"} Nov 24 08:12:37 crc kubenswrapper[4799]: I1124 08:12:37.588618 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"2c91ef74-c8d3-4283-913d-16cd52f50546","Type":"ContainerStarted","Data":"de50426cef9b6593ab335c1961195e0f322a829005fb9eac1313956387e3ac6f"} Nov 24 08:12:37 crc kubenswrapper[4799]: I1124 08:12:37.592030 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"b55c4e6c-8c8b-4623-b420-3cfb9eea5764","Type":"ContainerStarted","Data":"caeed878f93be152ad68e1562a3a6bfbd30e802c5334556d93eb2fe27b1f5f4f"} Nov 24 08:12:37 crc kubenswrapper[4799]: I1124 08:12:37.594162 4799 generic.go:334] "Generic (PLEG): container finished" podID="7f05f65a-f0e1-4645-8085-5614e23dc198" containerID="2a794cfe903d6ddfe89e23172c37df55f73ab3305a535a8ea7f5d0413d7a414b" exitCode=0 Nov 24 08:12:37 crc kubenswrapper[4799]: I1124 08:12:37.595068 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f05f65a-f0e1-4645-8085-5614e23dc198","Type":"ContainerDied","Data":"2a794cfe903d6ddfe89e23172c37df55f73ab3305a535a8ea7f5d0413d7a414b"} Nov 24 08:12:37 crc kubenswrapper[4799]: I1124 08:12:37.619559 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.290501087 podStartE2EDuration="2.619541548s" podCreationTimestamp="2025-11-24 08:12:35 +0000 UTC" firstStartedPulling="2025-11-24 08:12:36.402431957 +0000 UTC m=+5102.058414431" lastFinishedPulling="2025-11-24 08:12:36.731472408 +0000 UTC m=+5102.387454892" observedRunningTime="2025-11-24 08:12:37.608153425 +0000 UTC m=+5103.264135919" watchObservedRunningTime="2025-11-24 08:12:37.619541548 +0000 UTC m=+5103.275524022" Nov 24 08:12:37 crc kubenswrapper[4799]: I1124 08:12:37.634310 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.234799166 podStartE2EDuration="3.634287807s" podCreationTimestamp="2025-11-24 08:12:34 +0000 UTC" firstStartedPulling="2025-11-24 08:12:35.76459579 +0000 UTC m=+5101.420578264" lastFinishedPulling="2025-11-24 08:12:36.164084431 +0000 UTC m=+5101.820066905" observedRunningTime="2025-11-24 08:12:37.629687026 +0000 UTC m=+5103.285669500" watchObservedRunningTime="2025-11-24 08:12:37.634287807 +0000 UTC m=+5103.290270281" Nov 24 08:12:37 crc kubenswrapper[4799]: I1124 08:12:37.980026 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.110488 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjxk7\" (UniqueName: \"kubernetes.io/projected/7f05f65a-f0e1-4645-8085-5614e23dc198-kube-api-access-kjxk7\") pod \"7f05f65a-f0e1-4645-8085-5614e23dc198\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.110767 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f05f65a-f0e1-4645-8085-5614e23dc198-logs\") pod \"7f05f65a-f0e1-4645-8085-5614e23dc198\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.110803 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data-custom\") pod \"7f05f65a-f0e1-4645-8085-5614e23dc198\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.110876 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data\") pod \"7f05f65a-f0e1-4645-8085-5614e23dc198\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.110943 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-combined-ca-bundle\") pod \"7f05f65a-f0e1-4645-8085-5614e23dc198\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.111000 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-scripts\") pod \"7f05f65a-f0e1-4645-8085-5614e23dc198\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.111055 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f05f65a-f0e1-4645-8085-5614e23dc198-etc-machine-id\") pod \"7f05f65a-f0e1-4645-8085-5614e23dc198\" (UID: \"7f05f65a-f0e1-4645-8085-5614e23dc198\") " Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.111249 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f05f65a-f0e1-4645-8085-5614e23dc198-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7f05f65a-f0e1-4645-8085-5614e23dc198" (UID: "7f05f65a-f0e1-4645-8085-5614e23dc198"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.111717 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f05f65a-f0e1-4645-8085-5614e23dc198-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.111784 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f05f65a-f0e1-4645-8085-5614e23dc198-logs" (OuterVolumeSpecName: "logs") pod "7f05f65a-f0e1-4645-8085-5614e23dc198" (UID: "7f05f65a-f0e1-4645-8085-5614e23dc198"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.120960 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7f05f65a-f0e1-4645-8085-5614e23dc198" (UID: "7f05f65a-f0e1-4645-8085-5614e23dc198"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.121242 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f05f65a-f0e1-4645-8085-5614e23dc198-kube-api-access-kjxk7" (OuterVolumeSpecName: "kube-api-access-kjxk7") pod "7f05f65a-f0e1-4645-8085-5614e23dc198" (UID: "7f05f65a-f0e1-4645-8085-5614e23dc198"). InnerVolumeSpecName "kube-api-access-kjxk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.125380 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-scripts" (OuterVolumeSpecName: "scripts") pod "7f05f65a-f0e1-4645-8085-5614e23dc198" (UID: "7f05f65a-f0e1-4645-8085-5614e23dc198"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.151953 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f05f65a-f0e1-4645-8085-5614e23dc198" (UID: "7f05f65a-f0e1-4645-8085-5614e23dc198"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.184762 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data" (OuterVolumeSpecName: "config-data") pod "7f05f65a-f0e1-4645-8085-5614e23dc198" (UID: "7f05f65a-f0e1-4645-8085-5614e23dc198"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.213082 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f05f65a-f0e1-4645-8085-5614e23dc198-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.213127 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.213140 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.213148 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.213157 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f05f65a-f0e1-4645-8085-5614e23dc198-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.213165 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjxk7\" (UniqueName: \"kubernetes.io/projected/7f05f65a-f0e1-4645-8085-5614e23dc198-kube-api-access-kjxk7\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.595198 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.604283 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.605291 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f05f65a-f0e1-4645-8085-5614e23dc198","Type":"ContainerDied","Data":"d80b4683eb83a57e5ad16c4738a278bca3683a462c6107f11dc26c2500ff5934"} Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.605331 4799 scope.go:117] "RemoveContainer" containerID="2a794cfe903d6ddfe89e23172c37df55f73ab3305a535a8ea7f5d0413d7a414b" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.665356 4799 scope.go:117] "RemoveContainer" containerID="db78d862d8e4edd6078acf7c227757b2f6bd51963574eecb371a2dfe75e3b3bc" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.689720 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.727334 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.736551 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:12:38 crc kubenswrapper[4799]: E1124 08:12:38.736982 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f05f65a-f0e1-4645-8085-5614e23dc198" containerName="cinder-api" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.737000 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f05f65a-f0e1-4645-8085-5614e23dc198" containerName="cinder-api" Nov 24 08:12:38 crc kubenswrapper[4799]: E1124 08:12:38.737015 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f05f65a-f0e1-4645-8085-5614e23dc198" containerName="cinder-api-log" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.737021 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f05f65a-f0e1-4645-8085-5614e23dc198" containerName="cinder-api-log" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.737196 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f05f65a-f0e1-4645-8085-5614e23dc198" containerName="cinder-api-log" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.737210 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f05f65a-f0e1-4645-8085-5614e23dc198" containerName="cinder-api" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.738562 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.739820 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.748459 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.823123 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8eb0d76e-b94e-4120-aee5-395757ddd145-logs\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.823179 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.823209 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-config-data\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.823241 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8eb0d76e-b94e-4120-aee5-395757ddd145-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.823564 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-scripts\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.823643 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-config-data-custom\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.823754 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmsg4\" (UniqueName: \"kubernetes.io/projected/8eb0d76e-b94e-4120-aee5-395757ddd145-kube-api-access-pmsg4\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.922745 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.924710 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-scripts\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.924744 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-config-data-custom\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.924780 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmsg4\" (UniqueName: \"kubernetes.io/projected/8eb0d76e-b94e-4120-aee5-395757ddd145-kube-api-access-pmsg4\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.924823 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8eb0d76e-b94e-4120-aee5-395757ddd145-logs\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.924842 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.924880 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-config-data\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.924913 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8eb0d76e-b94e-4120-aee5-395757ddd145-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.925006 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8eb0d76e-b94e-4120-aee5-395757ddd145-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.925051 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.926044 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.926313 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8eb0d76e-b94e-4120-aee5-395757ddd145-logs\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.931702 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-config-data-custom\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.932478 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.933003 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-scripts\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.938974 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eb0d76e-b94e-4120-aee5-395757ddd145-config-data\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.941222 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.963656 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmsg4\" (UniqueName: \"kubernetes.io/projected/8eb0d76e-b94e-4120-aee5-395757ddd145-kube-api-access-pmsg4\") pod \"cinder-api-0\" (UID: \"8eb0d76e-b94e-4120-aee5-395757ddd145\") " pod="openstack/cinder-api-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.964806 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.969175 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 08:12:38 crc kubenswrapper[4799]: I1124 08:12:38.969272 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 08:12:39 crc kubenswrapper[4799]: I1124 08:12:39.065921 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:12:39 crc kubenswrapper[4799]: I1124 08:12:39.518917 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:12:39 crc kubenswrapper[4799]: W1124 08:12:39.526447 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8eb0d76e_b94e_4120_aee5_395757ddd145.slice/crio-f7c3bd2420d36c9adf973a268ba2ebefda165be510736d635a76257750945066 WatchSource:0}: Error finding container f7c3bd2420d36c9adf973a268ba2ebefda165be510736d635a76257750945066: Status 404 returned error can't find the container with id f7c3bd2420d36c9adf973a268ba2ebefda165be510736d635a76257750945066 Nov 24 08:12:39 crc kubenswrapper[4799]: I1124 08:12:39.622423 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8eb0d76e-b94e-4120-aee5-395757ddd145","Type":"ContainerStarted","Data":"f7c3bd2420d36c9adf973a268ba2ebefda165be510736d635a76257750945066"} Nov 24 08:12:39 crc kubenswrapper[4799]: I1124 08:12:39.624051 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 08:12:39 crc kubenswrapper[4799]: I1124 08:12:39.625962 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 08:12:39 crc kubenswrapper[4799]: I1124 08:12:39.655894 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f05f65a-f0e1-4645-8085-5614e23dc198" path="/var/lib/kubelet/pods/7f05f65a-f0e1-4645-8085-5614e23dc198/volumes" Nov 24 08:12:39 crc kubenswrapper[4799]: I1124 08:12:39.657388 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 08:12:40 crc kubenswrapper[4799]: I1124 08:12:40.218583 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:40 crc kubenswrapper[4799]: I1124 08:12:40.635929 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8eb0d76e-b94e-4120-aee5-395757ddd145","Type":"ContainerStarted","Data":"baf0de99d3d4a02a434bf87ec5d50a54745ab444b83cf1b3dfdb7d4f1fec6761"} Nov 24 08:12:40 crc kubenswrapper[4799]: I1124 08:12:40.815948 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Nov 24 08:12:41 crc kubenswrapper[4799]: I1124 08:12:41.651945 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8eb0d76e-b94e-4120-aee5-395757ddd145","Type":"ContainerStarted","Data":"d98c17b42f435ae5651433927e243e03c1a886161307705ac185ce979be53e81"} Nov 24 08:12:41 crc kubenswrapper[4799]: I1124 08:12:41.688534 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.688504768 podStartE2EDuration="3.688504768s" podCreationTimestamp="2025-11-24 08:12:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:12:41.676614771 +0000 UTC m=+5107.332597325" watchObservedRunningTime="2025-11-24 08:12:41.688504768 +0000 UTC m=+5107.344487252" Nov 24 08:12:42 crc kubenswrapper[4799]: I1124 08:12:42.664710 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 08:12:43 crc kubenswrapper[4799]: I1124 08:12:43.793980 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 08:12:43 crc kubenswrapper[4799]: I1124 08:12:43.862048 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:12:44 crc kubenswrapper[4799]: I1124 08:12:44.684347 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" containerName="cinder-scheduler" containerID="cri-o://53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf" gracePeriod=30 Nov 24 08:12:44 crc kubenswrapper[4799]: I1124 08:12:44.684423 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" containerName="probe" containerID="cri-o://4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad" gracePeriod=30 Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.417538 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.613530 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.682566 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data-custom\") pod \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.682629 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vqvk\" (UniqueName: \"kubernetes.io/projected/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-kube-api-access-6vqvk\") pod \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.682748 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-scripts\") pod \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.682816 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-etc-machine-id\") pod \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.682905 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-combined-ca-bundle\") pod \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.682945 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data\") pod \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\" (UID: \"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569\") " Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.684394 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" (UID: "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.691253 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-kube-api-access-6vqvk" (OuterVolumeSpecName: "kube-api-access-6vqvk") pod "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" (UID: "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569"). InnerVolumeSpecName "kube-api-access-6vqvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.707563 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-scripts" (OuterVolumeSpecName: "scripts") pod "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" (UID: "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.751140 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" (UID: "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.752306 4799 generic.go:334] "Generic (PLEG): container finished" podID="f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" containerID="4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad" exitCode=0 Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.752348 4799 generic.go:334] "Generic (PLEG): container finished" podID="f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" containerID="53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf" exitCode=0 Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.752375 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569","Type":"ContainerDied","Data":"4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad"} Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.752417 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569","Type":"ContainerDied","Data":"53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf"} Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.752429 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f6ef7571-ae8e-428b-b3a1-9fd1b6c95569","Type":"ContainerDied","Data":"046ff57d77d3f9c4530240fe1c12aeaeb91ca4346688447e15d6d5d2cad3ce9a"} Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.752450 4799 scope.go:117] "RemoveContainer" containerID="4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.752678 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.798818 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.798926 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vqvk\" (UniqueName: \"kubernetes.io/projected/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-kube-api-access-6vqvk\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.798939 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.798950 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.830242 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" (UID: "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.839162 4799 scope.go:117] "RemoveContainer" containerID="53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.876665 4799 scope.go:117] "RemoveContainer" containerID="4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad" Nov 24 08:12:45 crc kubenswrapper[4799]: E1124 08:12:45.877168 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad\": container with ID starting with 4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad not found: ID does not exist" containerID="4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.877197 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad"} err="failed to get container status \"4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad\": rpc error: code = NotFound desc = could not find container \"4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad\": container with ID starting with 4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad not found: ID does not exist" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.877220 4799 scope.go:117] "RemoveContainer" containerID="53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf" Nov 24 08:12:45 crc kubenswrapper[4799]: E1124 08:12:45.877605 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf\": container with ID starting with 53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf not found: ID does not exist" containerID="53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.877652 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf"} err="failed to get container status \"53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf\": rpc error: code = NotFound desc = could not find container \"53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf\": container with ID starting with 53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf not found: ID does not exist" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.877683 4799 scope.go:117] "RemoveContainer" containerID="4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.878475 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad"} err="failed to get container status \"4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad\": rpc error: code = NotFound desc = could not find container \"4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad\": container with ID starting with 4e49035b2c00d4996108e0660e8ee4a90cdae13ea4ddf59ddff8e4904f3900ad not found: ID does not exist" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.878510 4799 scope.go:117] "RemoveContainer" containerID="53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.879730 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf"} err="failed to get container status \"53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf\": rpc error: code = NotFound desc = could not find container \"53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf\": container with ID starting with 53202b30ea4ee9fc5651905f6283d110e0717d4bb5f8f0e03d4b7b56ba3cb9cf not found: ID does not exist" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.900966 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:45 crc kubenswrapper[4799]: I1124 08:12:45.915833 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data" (OuterVolumeSpecName: "config-data") pod "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" (UID: "f6ef7571-ae8e-428b-b3a1-9fd1b6c95569"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.003164 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.058711 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.136510 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.148908 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.161140 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:12:46 crc kubenswrapper[4799]: E1124 08:12:46.161601 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" containerName="probe" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.161624 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" containerName="probe" Nov 24 08:12:46 crc kubenswrapper[4799]: E1124 08:12:46.161643 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" containerName="cinder-scheduler" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.161652 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" containerName="cinder-scheduler" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.161905 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" containerName="cinder-scheduler" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.161948 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" containerName="probe" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.164345 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.171865 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.178897 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.309664 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-config-data\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.309775 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.309817 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-scripts\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.309894 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.309932 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c14bf25c-677c-493a-a8d5-dec06d445bb8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.309951 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4rg4\" (UniqueName: \"kubernetes.io/projected/c14bf25c-677c-493a-a8d5-dec06d445bb8-kube-api-access-x4rg4\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.411772 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.411941 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-scripts\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.412018 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.412065 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c14bf25c-677c-493a-a8d5-dec06d445bb8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.412087 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4rg4\" (UniqueName: \"kubernetes.io/projected/c14bf25c-677c-493a-a8d5-dec06d445bb8-kube-api-access-x4rg4\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.412154 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-config-data\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.412574 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c14bf25c-677c-493a-a8d5-dec06d445bb8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.418444 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-scripts\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.418801 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.420277 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.434928 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14bf25c-677c-493a-a8d5-dec06d445bb8-config-data\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.442439 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4rg4\" (UniqueName: \"kubernetes.io/projected/c14bf25c-677c-493a-a8d5-dec06d445bb8-kube-api-access-x4rg4\") pod \"cinder-scheduler-0\" (UID: \"c14bf25c-677c-493a-a8d5-dec06d445bb8\") " pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.488255 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.629233 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:12:46 crc kubenswrapper[4799]: E1124 08:12:46.629806 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:12:46 crc kubenswrapper[4799]: I1124 08:12:46.975126 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:12:47 crc kubenswrapper[4799]: I1124 08:12:47.642149 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6ef7571-ae8e-428b-b3a1-9fd1b6c95569" path="/var/lib/kubelet/pods/f6ef7571-ae8e-428b-b3a1-9fd1b6c95569/volumes" Nov 24 08:12:47 crc kubenswrapper[4799]: I1124 08:12:47.774518 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c14bf25c-677c-493a-a8d5-dec06d445bb8","Type":"ContainerStarted","Data":"ea62e28b807fb761e3901a2517d97542a644e76d7b61266a9327f6db55481575"} Nov 24 08:12:47 crc kubenswrapper[4799]: I1124 08:12:47.774569 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c14bf25c-677c-493a-a8d5-dec06d445bb8","Type":"ContainerStarted","Data":"2ac2c0dfcc50549b5c2a36f4597877ce52b4533f4c3f5492afe48bc3870a285e"} Nov 24 08:12:48 crc kubenswrapper[4799]: I1124 08:12:48.788266 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c14bf25c-677c-493a-a8d5-dec06d445bb8","Type":"ContainerStarted","Data":"c7ef413af249b9e3bb873e64daab69402bd7cf334407aec8f38818461c930a34"} Nov 24 08:12:48 crc kubenswrapper[4799]: I1124 08:12:48.823024 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.822998533 podStartE2EDuration="2.822998533s" podCreationTimestamp="2025-11-24 08:12:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:12:48.812827964 +0000 UTC m=+5114.468810438" watchObservedRunningTime="2025-11-24 08:12:48.822998533 +0000 UTC m=+5114.478981017" Nov 24 08:12:50 crc kubenswrapper[4799]: I1124 08:12:50.914999 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 08:12:51 crc kubenswrapper[4799]: I1124 08:12:51.489750 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 08:12:56 crc kubenswrapper[4799]: I1124 08:12:56.744555 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 08:12:59 crc kubenswrapper[4799]: I1124 08:12:59.628398 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:12:59 crc kubenswrapper[4799]: E1124 08:12:59.628923 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:13:13 crc kubenswrapper[4799]: I1124 08:13:13.628506 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:13:13 crc kubenswrapper[4799]: E1124 08:13:13.629303 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:13:25 crc kubenswrapper[4799]: I1124 08:13:25.634432 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:13:25 crc kubenswrapper[4799]: E1124 08:13:25.635255 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:13:33 crc kubenswrapper[4799]: I1124 08:13:33.056356 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-4241-account-create-vcqfd"] Nov 24 08:13:33 crc kubenswrapper[4799]: I1124 08:13:33.066482 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-4d9lb"] Nov 24 08:13:33 crc kubenswrapper[4799]: I1124 08:13:33.074736 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-4d9lb"] Nov 24 08:13:33 crc kubenswrapper[4799]: I1124 08:13:33.103341 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-4241-account-create-vcqfd"] Nov 24 08:13:33 crc kubenswrapper[4799]: I1124 08:13:33.638616 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3" path="/var/lib/kubelet/pods/b5c93b3f-1bca-4c1a-90ff-5e8bc896dcf3/volumes" Nov 24 08:13:33 crc kubenswrapper[4799]: I1124 08:13:33.639516 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbbb348f-6084-4a19-81a6-f3e5027df7fb" path="/var/lib/kubelet/pods/bbbb348f-6084-4a19-81a6-f3e5027df7fb/volumes" Nov 24 08:13:40 crc kubenswrapper[4799]: I1124 08:13:40.628341 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:13:40 crc kubenswrapper[4799]: E1124 08:13:40.629182 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:13:45 crc kubenswrapper[4799]: I1124 08:13:45.040155 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-56drr"] Nov 24 08:13:45 crc kubenswrapper[4799]: I1124 08:13:45.055843 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-56drr"] Nov 24 08:13:45 crc kubenswrapper[4799]: I1124 08:13:45.649176 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ba86795-13a5-4f7b-bc59-07a23af8b91f" path="/var/lib/kubelet/pods/7ba86795-13a5-4f7b-bc59-07a23af8b91f/volumes" Nov 24 08:13:54 crc kubenswrapper[4799]: I1124 08:13:54.628493 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:13:54 crc kubenswrapper[4799]: E1124 08:13:54.629399 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:13:59 crc kubenswrapper[4799]: I1124 08:13:59.042977 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-p56f8"] Nov 24 08:13:59 crc kubenswrapper[4799]: I1124 08:13:59.051849 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-p56f8"] Nov 24 08:13:59 crc kubenswrapper[4799]: I1124 08:13:59.535552 4799 scope.go:117] "RemoveContainer" containerID="17595c10319016d861b3590599c798d01b512c7bd2120caecf538b9b3f03311b" Nov 24 08:13:59 crc kubenswrapper[4799]: I1124 08:13:59.575092 4799 scope.go:117] "RemoveContainer" containerID="58f943f677764a5d1dff6bc7526a87ad6ee6c30c8c762e248f2a042890f217e2" Nov 24 08:13:59 crc kubenswrapper[4799]: I1124 08:13:59.644434 4799 scope.go:117] "RemoveContainer" containerID="75b1b31f8be72887161c747495a6c8bbd55f3a10a928ef86b2e15c85759f47c7" Nov 24 08:13:59 crc kubenswrapper[4799]: I1124 08:13:59.649184 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3df32d62-80d5-47ab-a81c-1ce3818eb9be" path="/var/lib/kubelet/pods/3df32d62-80d5-47ab-a81c-1ce3818eb9be/volumes" Nov 24 08:13:59 crc kubenswrapper[4799]: I1124 08:13:59.686172 4799 scope.go:117] "RemoveContainer" containerID="7b14e2ac73ede37055cbdf19b9c152c448f6cebad58d91196b5539fa13d4d253" Nov 24 08:14:07 crc kubenswrapper[4799]: I1124 08:14:07.628504 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:14:07 crc kubenswrapper[4799]: E1124 08:14:07.629709 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:14:19 crc kubenswrapper[4799]: I1124 08:14:19.628806 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:14:19 crc kubenswrapper[4799]: E1124 08:14:19.629670 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.022067 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5bccb8fd89-sll6f"] Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.028418 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.032625 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.032815 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-s5jpd" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.033224 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.033344 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.058033 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bccb8fd89-sll6f"] Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.100994 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.101245 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" containerName="glance-log" containerID="cri-o://dc14a6488fdb8f0b78a111c5df96c45942e6965d8020be92ae627a6559759dac" gracePeriod=30 Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.101637 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" containerName="glance-httpd" containerID="cri-o://3b8326afd0a897e5ef200985d1454222061c289ca54efbb6114e721972fbdeed" gracePeriod=30 Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.145698 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-57d54fb7f-xkrjn"] Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.147279 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.163708 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-config-data\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.163747 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b85335a2-d197-47b5-be68-e45201d3657f-logs\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.163775 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-scripts\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.163988 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tlq4\" (UniqueName: \"kubernetes.io/projected/b85335a2-d197-47b5-be68-e45201d3657f-kube-api-access-9tlq4\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.164170 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b85335a2-d197-47b5-be68-e45201d3657f-horizon-secret-key\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.180463 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-57d54fb7f-xkrjn"] Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.218267 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.218602 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" containerName="glance-log" containerID="cri-o://c1939c2300d3acc7a7b5dcf826e0a316e19046ae2e5527f9484594291f376394" gracePeriod=30 Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.219288 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" containerName="glance-httpd" containerID="cri-o://9baa784af9346bce45134500375430672747e10413316507b7bed0357ee4d7d8" gracePeriod=30 Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.265643 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e498b0a-27e6-4e14-9440-48e18e327e70-horizon-secret-key\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.265708 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-config-data\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.265729 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b85335a2-d197-47b5-be68-e45201d3657f-logs\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.265754 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-scripts\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.265774 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-config-data\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.265807 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-scripts\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.265834 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tlq4\" (UniqueName: \"kubernetes.io/projected/b85335a2-d197-47b5-be68-e45201d3657f-kube-api-access-9tlq4\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.266013 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwdbm\" (UniqueName: \"kubernetes.io/projected/2e498b0a-27e6-4e14-9440-48e18e327e70-kube-api-access-xwdbm\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.266098 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b85335a2-d197-47b5-be68-e45201d3657f-horizon-secret-key\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.266165 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e498b0a-27e6-4e14-9440-48e18e327e70-logs\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.266293 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b85335a2-d197-47b5-be68-e45201d3657f-logs\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.266521 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-scripts\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.267145 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-config-data\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.271943 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b85335a2-d197-47b5-be68-e45201d3657f-horizon-secret-key\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.284971 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tlq4\" (UniqueName: \"kubernetes.io/projected/b85335a2-d197-47b5-be68-e45201d3657f-kube-api-access-9tlq4\") pod \"horizon-5bccb8fd89-sll6f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.358531 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.367414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-config-data\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.367478 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-scripts\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.367552 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwdbm\" (UniqueName: \"kubernetes.io/projected/2e498b0a-27e6-4e14-9440-48e18e327e70-kube-api-access-xwdbm\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.367613 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e498b0a-27e6-4e14-9440-48e18e327e70-logs\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.367683 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e498b0a-27e6-4e14-9440-48e18e327e70-horizon-secret-key\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.368248 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e498b0a-27e6-4e14-9440-48e18e327e70-logs\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.368647 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-scripts\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.369093 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-config-data\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.371096 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e498b0a-27e6-4e14-9440-48e18e327e70-horizon-secret-key\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.387892 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwdbm\" (UniqueName: \"kubernetes.io/projected/2e498b0a-27e6-4e14-9440-48e18e327e70-kube-api-access-xwdbm\") pod \"horizon-57d54fb7f-xkrjn\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.473401 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.715563 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bccb8fd89-sll6f"] Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.767245 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6896b5c567-jgd8d"] Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.769361 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.791405 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6896b5c567-jgd8d"] Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.825595 4799 generic.go:334] "Generic (PLEG): container finished" podID="363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" containerID="dc14a6488fdb8f0b78a111c5df96c45942e6965d8020be92ae627a6559759dac" exitCode=143 Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.825655 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42","Type":"ContainerDied","Data":"dc14a6488fdb8f0b78a111c5df96c45942e6965d8020be92ae627a6559759dac"} Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.828985 4799 generic.go:334] "Generic (PLEG): container finished" podID="46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" containerID="c1939c2300d3acc7a7b5dcf826e0a316e19046ae2e5527f9484594291f376394" exitCode=143 Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.829015 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949","Type":"ContainerDied","Data":"c1939c2300d3acc7a7b5dcf826e0a316e19046ae2e5527f9484594291f376394"} Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.834907 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bccb8fd89-sll6f"] Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.849523 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.886939 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-scripts\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.887018 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-config-data\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.887077 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9wj6\" (UniqueName: \"kubernetes.io/projected/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-kube-api-access-w9wj6\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.887107 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-logs\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.887188 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-horizon-secret-key\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.988809 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-scripts\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.988882 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-config-data\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.988922 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9wj6\" (UniqueName: \"kubernetes.io/projected/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-kube-api-access-w9wj6\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.988945 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-logs\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.988966 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-horizon-secret-key\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.990297 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-logs\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.991155 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-config-data\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.991641 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-scripts\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:29 crc kubenswrapper[4799]: I1124 08:14:29.993665 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-horizon-secret-key\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:30 crc kubenswrapper[4799]: I1124 08:14:30.009036 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9wj6\" (UniqueName: \"kubernetes.io/projected/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-kube-api-access-w9wj6\") pod \"horizon-6896b5c567-jgd8d\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:30 crc kubenswrapper[4799]: I1124 08:14:30.102717 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-57d54fb7f-xkrjn"] Nov 24 08:14:30 crc kubenswrapper[4799]: I1124 08:14:30.103164 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:30 crc kubenswrapper[4799]: W1124 08:14:30.105411 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e498b0a_27e6_4e14_9440_48e18e327e70.slice/crio-604f80a945ffa378752cd30a785981143298632516d586f3549a32cb58764cc3 WatchSource:0}: Error finding container 604f80a945ffa378752cd30a785981143298632516d586f3549a32cb58764cc3: Status 404 returned error can't find the container with id 604f80a945ffa378752cd30a785981143298632516d586f3549a32cb58764cc3 Nov 24 08:14:30 crc kubenswrapper[4799]: I1124 08:14:30.566966 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6896b5c567-jgd8d"] Nov 24 08:14:30 crc kubenswrapper[4799]: I1124 08:14:30.845286 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6896b5c567-jgd8d" event={"ID":"700b8e45-d105-4ca5-b89c-0c7c0171c0fb","Type":"ContainerStarted","Data":"9462e48835f5c7546f11e23bc846fe7531de5b26b9697a42aaaa91429451e0f4"} Nov 24 08:14:30 crc kubenswrapper[4799]: I1124 08:14:30.848136 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d54fb7f-xkrjn" event={"ID":"2e498b0a-27e6-4e14-9440-48e18e327e70","Type":"ContainerStarted","Data":"604f80a945ffa378752cd30a785981143298632516d586f3549a32cb58764cc3"} Nov 24 08:14:30 crc kubenswrapper[4799]: I1124 08:14:30.849635 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bccb8fd89-sll6f" event={"ID":"b85335a2-d197-47b5-be68-e45201d3657f","Type":"ContainerStarted","Data":"33b44145a34e2a77449221b5fb927d7043b856d2fe49a74291ce962b193c68d7"} Nov 24 08:14:31 crc kubenswrapper[4799]: I1124 08:14:31.629226 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:14:31 crc kubenswrapper[4799]: E1124 08:14:31.629783 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:14:32 crc kubenswrapper[4799]: I1124 08:14:32.872989 4799 generic.go:334] "Generic (PLEG): container finished" podID="363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" containerID="3b8326afd0a897e5ef200985d1454222061c289ca54efbb6114e721972fbdeed" exitCode=0 Nov 24 08:14:32 crc kubenswrapper[4799]: I1124 08:14:32.873052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42","Type":"ContainerDied","Data":"3b8326afd0a897e5ef200985d1454222061c289ca54efbb6114e721972fbdeed"} Nov 24 08:14:32 crc kubenswrapper[4799]: I1124 08:14:32.873080 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42","Type":"ContainerDied","Data":"3c2f8f15e9b9b6f8f29b3cccdd0124a600ab7aa8030cbb00c90952409b212812"} Nov 24 08:14:32 crc kubenswrapper[4799]: I1124 08:14:32.873091 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c2f8f15e9b9b6f8f29b3cccdd0124a600ab7aa8030cbb00c90952409b212812" Nov 24 08:14:32 crc kubenswrapper[4799]: I1124 08:14:32.875124 4799 generic.go:334] "Generic (PLEG): container finished" podID="46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" containerID="9baa784af9346bce45134500375430672747e10413316507b7bed0357ee4d7d8" exitCode=0 Nov 24 08:14:32 crc kubenswrapper[4799]: I1124 08:14:32.875145 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949","Type":"ContainerDied","Data":"9baa784af9346bce45134500375430672747e10413316507b7bed0357ee4d7d8"} Nov 24 08:14:32 crc kubenswrapper[4799]: I1124 08:14:32.875159 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949","Type":"ContainerDied","Data":"0e71411b4c7933effbdf63bfe8b4f4b18ecbab01ac7bf9ce362c276d9f96dd2e"} Nov 24 08:14:32 crc kubenswrapper[4799]: I1124 08:14:32.875169 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e71411b4c7933effbdf63bfe8b4f4b18ecbab01ac7bf9ce362c276d9f96dd2e" Nov 24 08:14:32 crc kubenswrapper[4799]: I1124 08:14:32.936452 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 08:14:32 crc kubenswrapper[4799]: I1124 08:14:32.941780 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.055806 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-scripts\") pod \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.055947 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-combined-ca-bundle\") pod \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.055995 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-ceph\") pod \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.056024 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slkkn\" (UniqueName: \"kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-kube-api-access-slkkn\") pod \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.056055 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-logs\") pod \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.056098 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvx8q\" (UniqueName: \"kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-kube-api-access-jvx8q\") pod \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.056165 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-ceph\") pod \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.056244 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-logs\") pod \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.056273 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-httpd-run\") pod \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.056299 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-config-data\") pod \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.056321 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-scripts\") pod \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.056362 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-httpd-run\") pod \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.056388 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-combined-ca-bundle\") pod \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\" (UID: \"46b45ccf-b3d0-44da-a16e-8d1c5a8c3949\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.056462 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-config-data\") pod \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\" (UID: \"363dee68-bcb4-45f0-a16c-ab6a4d0d2b42\") " Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.062348 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-logs" (OuterVolumeSpecName: "logs") pod "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" (UID: "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.062363 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" (UID: "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.062414 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" (UID: "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.062902 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-logs" (OuterVolumeSpecName: "logs") pod "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" (UID: "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.063488 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-ceph" (OuterVolumeSpecName: "ceph") pod "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" (UID: "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.064010 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-ceph" (OuterVolumeSpecName: "ceph") pod "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" (UID: "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.064995 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-scripts" (OuterVolumeSpecName: "scripts") pod "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" (UID: "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.065289 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-kube-api-access-jvx8q" (OuterVolumeSpecName: "kube-api-access-jvx8q") pod "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" (UID: "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42"). InnerVolumeSpecName "kube-api-access-jvx8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.066026 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-kube-api-access-slkkn" (OuterVolumeSpecName: "kube-api-access-slkkn") pod "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" (UID: "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949"). InnerVolumeSpecName "kube-api-access-slkkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.073080 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-scripts" (OuterVolumeSpecName: "scripts") pod "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" (UID: "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.096158 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" (UID: "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.096884 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" (UID: "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.117212 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-config-data" (OuterVolumeSpecName: "config-data") pod "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" (UID: "363dee68-bcb4-45f0-a16c-ab6a4d0d2b42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.125313 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-config-data" (OuterVolumeSpecName: "config-data") pod "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" (UID: "46b45ccf-b3d0-44da-a16e-8d1c5a8c3949"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158468 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158507 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158520 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158531 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158545 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158556 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slkkn\" (UniqueName: \"kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-kube-api-access-slkkn\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158567 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158578 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvx8q\" (UniqueName: \"kubernetes.io/projected/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-kube-api-access-jvx8q\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158589 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158599 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158609 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158621 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158633 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.158643 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.882385 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.882440 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.908963 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.918286 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.939089 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:14:33 crc kubenswrapper[4799]: E1124 08:14:33.939469 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" containerName="glance-log" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.939481 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" containerName="glance-log" Nov 24 08:14:33 crc kubenswrapper[4799]: E1124 08:14:33.939499 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" containerName="glance-httpd" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.939505 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" containerName="glance-httpd" Nov 24 08:14:33 crc kubenswrapper[4799]: E1124 08:14:33.939526 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" containerName="glance-httpd" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.939532 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" containerName="glance-httpd" Nov 24 08:14:33 crc kubenswrapper[4799]: E1124 08:14:33.939543 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" containerName="glance-log" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.939549 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" containerName="glance-log" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.939703 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" containerName="glance-log" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.939719 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" containerName="glance-httpd" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.939734 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" containerName="glance-httpd" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.939751 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" containerName="glance-log" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.940705 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.943685 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.943839 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lwt7f" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.944340 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.948370 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.958653 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.964084 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.970760 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.972236 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:33 crc kubenswrapper[4799]: I1124 08:14:33.973876 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.000598 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.088950 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4a7265-94b0-4631-9500-4ca4e8e4af26-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.088996 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b86h\" (UniqueName: \"kubernetes.io/projected/ea055bb9-dd38-48ef-a248-83b68e3176b6-kube-api-access-9b86h\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089025 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7f4a7265-94b0-4631-9500-4ca4e8e4af26-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089045 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f4a7265-94b0-4631-9500-4ca4e8e4af26-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089069 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea055bb9-dd38-48ef-a248-83b68e3176b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089134 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f4a7265-94b0-4631-9500-4ca4e8e4af26-logs\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089181 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea055bb9-dd38-48ef-a248-83b68e3176b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089208 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4a7265-94b0-4631-9500-4ca4e8e4af26-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089291 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ea055bb9-dd38-48ef-a248-83b68e3176b6-ceph\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089319 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea055bb9-dd38-48ef-a248-83b68e3176b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089417 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea055bb9-dd38-48ef-a248-83b68e3176b6-logs\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089469 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea055bb9-dd38-48ef-a248-83b68e3176b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089495 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f4a7265-94b0-4631-9500-4ca4e8e4af26-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.089525 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgrhg\" (UniqueName: \"kubernetes.io/projected/7f4a7265-94b0-4631-9500-4ca4e8e4af26-kube-api-access-hgrhg\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.190907 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea055bb9-dd38-48ef-a248-83b68e3176b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.190955 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f4a7265-94b0-4631-9500-4ca4e8e4af26-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.190978 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgrhg\" (UniqueName: \"kubernetes.io/projected/7f4a7265-94b0-4631-9500-4ca4e8e4af26-kube-api-access-hgrhg\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.191012 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4a7265-94b0-4631-9500-4ca4e8e4af26-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.191033 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b86h\" (UniqueName: \"kubernetes.io/projected/ea055bb9-dd38-48ef-a248-83b68e3176b6-kube-api-access-9b86h\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.191057 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7f4a7265-94b0-4631-9500-4ca4e8e4af26-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.191076 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f4a7265-94b0-4631-9500-4ca4e8e4af26-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.191103 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea055bb9-dd38-48ef-a248-83b68e3176b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.191125 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f4a7265-94b0-4631-9500-4ca4e8e4af26-logs\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.191157 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea055bb9-dd38-48ef-a248-83b68e3176b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.191177 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4a7265-94b0-4631-9500-4ca4e8e4af26-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.191206 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ea055bb9-dd38-48ef-a248-83b68e3176b6-ceph\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.191223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea055bb9-dd38-48ef-a248-83b68e3176b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.191284 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea055bb9-dd38-48ef-a248-83b68e3176b6-logs\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.192459 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f4a7265-94b0-4631-9500-4ca4e8e4af26-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.193589 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea055bb9-dd38-48ef-a248-83b68e3176b6-logs\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.203545 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f4a7265-94b0-4631-9500-4ca4e8e4af26-logs\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.203943 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea055bb9-dd38-48ef-a248-83b68e3176b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.204198 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea055bb9-dd38-48ef-a248-83b68e3176b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.205261 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea055bb9-dd38-48ef-a248-83b68e3176b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.211209 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4a7265-94b0-4631-9500-4ca4e8e4af26-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.211843 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7f4a7265-94b0-4631-9500-4ca4e8e4af26-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.211991 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f4a7265-94b0-4631-9500-4ca4e8e4af26-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.211895 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ea055bb9-dd38-48ef-a248-83b68e3176b6-ceph\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.214701 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgrhg\" (UniqueName: \"kubernetes.io/projected/7f4a7265-94b0-4631-9500-4ca4e8e4af26-kube-api-access-hgrhg\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.215212 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea055bb9-dd38-48ef-a248-83b68e3176b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.216795 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b86h\" (UniqueName: \"kubernetes.io/projected/ea055bb9-dd38-48ef-a248-83b68e3176b6-kube-api-access-9b86h\") pod \"glance-default-external-api-0\" (UID: \"ea055bb9-dd38-48ef-a248-83b68e3176b6\") " pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.236519 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4a7265-94b0-4631-9500-4ca4e8e4af26-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7f4a7265-94b0-4631-9500-4ca4e8e4af26\") " pod="openstack/glance-default-internal-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.307918 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 08:14:34 crc kubenswrapper[4799]: I1124 08:14:34.317164 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:35 crc kubenswrapper[4799]: I1124 08:14:35.644475 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="363dee68-bcb4-45f0-a16c-ab6a4d0d2b42" path="/var/lib/kubelet/pods/363dee68-bcb4-45f0-a16c-ab6a4d0d2b42/volumes" Nov 24 08:14:35 crc kubenswrapper[4799]: I1124 08:14:35.645439 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46b45ccf-b3d0-44da-a16e-8d1c5a8c3949" path="/var/lib/kubelet/pods/46b45ccf-b3d0-44da-a16e-8d1c5a8c3949/volumes" Nov 24 08:14:37 crc kubenswrapper[4799]: I1124 08:14:37.866334 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nc5j2"] Nov 24 08:14:37 crc kubenswrapper[4799]: I1124 08:14:37.880272 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nc5j2"] Nov 24 08:14:37 crc kubenswrapper[4799]: I1124 08:14:37.880395 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:37 crc kubenswrapper[4799]: I1124 08:14:37.967893 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-catalog-content\") pod \"community-operators-nc5j2\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:37 crc kubenswrapper[4799]: I1124 08:14:37.968128 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsg5j\" (UniqueName: \"kubernetes.io/projected/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-kube-api-access-jsg5j\") pod \"community-operators-nc5j2\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:37 crc kubenswrapper[4799]: I1124 08:14:37.968215 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-utilities\") pod \"community-operators-nc5j2\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.071977 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-utilities\") pod \"community-operators-nc5j2\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.072069 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-catalog-content\") pod \"community-operators-nc5j2\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.072115 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsg5j\" (UniqueName: \"kubernetes.io/projected/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-kube-api-access-jsg5j\") pod \"community-operators-nc5j2\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.072817 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-utilities\") pod \"community-operators-nc5j2\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.074489 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-catalog-content\") pod \"community-operators-nc5j2\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.089974 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsg5j\" (UniqueName: \"kubernetes.io/projected/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-kube-api-access-jsg5j\") pod \"community-operators-nc5j2\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.207123 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.559049 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.669144 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.892287 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nc5j2"] Nov 24 08:14:38 crc kubenswrapper[4799]: W1124 08:14:38.897535 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbbf0c9d8_5382_4ccc_b433_c63cd5cf7acd.slice/crio-da77073b56b1e623cab0b0a66d4e3eadbdb8ee5ffbac3d55666cc190bc177886 WatchSource:0}: Error finding container da77073b56b1e623cab0b0a66d4e3eadbdb8ee5ffbac3d55666cc190bc177886: Status 404 returned error can't find the container with id da77073b56b1e623cab0b0a66d4e3eadbdb8ee5ffbac3d55666cc190bc177886 Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.938959 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d54fb7f-xkrjn" event={"ID":"2e498b0a-27e6-4e14-9440-48e18e327e70","Type":"ContainerStarted","Data":"a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd"} Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.939015 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d54fb7f-xkrjn" event={"ID":"2e498b0a-27e6-4e14-9440-48e18e327e70","Type":"ContainerStarted","Data":"3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c"} Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.941389 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bccb8fd89-sll6f" event={"ID":"b85335a2-d197-47b5-be68-e45201d3657f","Type":"ContainerStarted","Data":"d009d073bd9d809d5a3bf3a99ec77b9e48ca66e7bc1fcf6cf11932881e2b264e"} Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.941416 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bccb8fd89-sll6f" event={"ID":"b85335a2-d197-47b5-be68-e45201d3657f","Type":"ContainerStarted","Data":"7eff771f603f594f1c891f04bffed0b19dce8367713050adf97f181b9a2ecf49"} Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.941519 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5bccb8fd89-sll6f" podUID="b85335a2-d197-47b5-be68-e45201d3657f" containerName="horizon-log" containerID="cri-o://7eff771f603f594f1c891f04bffed0b19dce8367713050adf97f181b9a2ecf49" gracePeriod=30 Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.941803 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5bccb8fd89-sll6f" podUID="b85335a2-d197-47b5-be68-e45201d3657f" containerName="horizon" containerID="cri-o://d009d073bd9d809d5a3bf3a99ec77b9e48ca66e7bc1fcf6cf11932881e2b264e" gracePeriod=30 Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.943832 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nc5j2" event={"ID":"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd","Type":"ContainerStarted","Data":"da77073b56b1e623cab0b0a66d4e3eadbdb8ee5ffbac3d55666cc190bc177886"} Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.944720 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ea055bb9-dd38-48ef-a248-83b68e3176b6","Type":"ContainerStarted","Data":"71b51787f523c41d54c4f8b38a76840de7f982918c9abae01ed6eb336861524d"} Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.945770 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7f4a7265-94b0-4631-9500-4ca4e8e4af26","Type":"ContainerStarted","Data":"6d2b22d0647a8cd6e7c8e7c25e55d5b2a271b89a7a16b5cc77a59ae629202027"} Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.947816 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6896b5c567-jgd8d" event={"ID":"700b8e45-d105-4ca5-b89c-0c7c0171c0fb","Type":"ContainerStarted","Data":"82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c"} Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.947858 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6896b5c567-jgd8d" event={"ID":"700b8e45-d105-4ca5-b89c-0c7c0171c0fb","Type":"ContainerStarted","Data":"71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744"} Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.990275 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-57d54fb7f-xkrjn" podStartSLOduration=2.11373196 podStartE2EDuration="9.99025554s" podCreationTimestamp="2025-11-24 08:14:29 +0000 UTC" firstStartedPulling="2025-11-24 08:14:30.125216419 +0000 UTC m=+5215.781198903" lastFinishedPulling="2025-11-24 08:14:38.001740009 +0000 UTC m=+5223.657722483" observedRunningTime="2025-11-24 08:14:38.966460915 +0000 UTC m=+5224.622443389" watchObservedRunningTime="2025-11-24 08:14:38.99025554 +0000 UTC m=+5224.646238014" Nov 24 08:14:38 crc kubenswrapper[4799]: I1124 08:14:38.994112 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5bccb8fd89-sll6f" podStartSLOduration=2.799036546 podStartE2EDuration="10.994098579s" podCreationTimestamp="2025-11-24 08:14:28 +0000 UTC" firstStartedPulling="2025-11-24 08:14:29.849352138 +0000 UTC m=+5215.505334612" lastFinishedPulling="2025-11-24 08:14:38.044414171 +0000 UTC m=+5223.700396645" observedRunningTime="2025-11-24 08:14:38.987374839 +0000 UTC m=+5224.643357313" watchObservedRunningTime="2025-11-24 08:14:38.994098579 +0000 UTC m=+5224.650081053" Nov 24 08:14:39 crc kubenswrapper[4799]: I1124 08:14:39.019534 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6896b5c567-jgd8d" podStartSLOduration=2.586971065 podStartE2EDuration="10.019514831s" podCreationTimestamp="2025-11-24 08:14:29 +0000 UTC" firstStartedPulling="2025-11-24 08:14:30.590248641 +0000 UTC m=+5216.246231115" lastFinishedPulling="2025-11-24 08:14:38.022792407 +0000 UTC m=+5223.678774881" observedRunningTime="2025-11-24 08:14:39.006504472 +0000 UTC m=+5224.662486946" watchObservedRunningTime="2025-11-24 08:14:39.019514831 +0000 UTC m=+5224.675497305" Nov 24 08:14:39 crc kubenswrapper[4799]: I1124 08:14:39.359582 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:14:39 crc kubenswrapper[4799]: I1124 08:14:39.702973 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:39 crc kubenswrapper[4799]: I1124 08:14:39.705552 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:14:39 crc kubenswrapper[4799]: I1124 08:14:39.979930 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ea055bb9-dd38-48ef-a248-83b68e3176b6","Type":"ContainerStarted","Data":"a429b7de71ac32d7f3ec7934d761a8aef894c91d547833f23544d691555d1115"} Nov 24 08:14:39 crc kubenswrapper[4799]: I1124 08:14:39.979976 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ea055bb9-dd38-48ef-a248-83b68e3176b6","Type":"ContainerStarted","Data":"ba4954cfc946e15439d20edd4b7b0fdd2b2b016e5542ae0afd0bf5247ef25593"} Nov 24 08:14:39 crc kubenswrapper[4799]: I1124 08:14:39.984147 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7f4a7265-94b0-4631-9500-4ca4e8e4af26","Type":"ContainerStarted","Data":"3cefef58ca1ffbcb23fc525297d02855f68796bebb38975b1fa7fa31b70d6da2"} Nov 24 08:14:39 crc kubenswrapper[4799]: I1124 08:14:39.985971 4799 generic.go:334] "Generic (PLEG): container finished" podID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" containerID="cec464f06c7c8cabbc0cfe7babefc732b668decafba04d972e1eb5d1258422ee" exitCode=0 Nov 24 08:14:39 crc kubenswrapper[4799]: I1124 08:14:39.987052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nc5j2" event={"ID":"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd","Type":"ContainerDied","Data":"cec464f06c7c8cabbc0cfe7babefc732b668decafba04d972e1eb5d1258422ee"} Nov 24 08:14:40 crc kubenswrapper[4799]: I1124 08:14:40.003209 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.003170515 podStartE2EDuration="7.003170515s" podCreationTimestamp="2025-11-24 08:14:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:14:40.001326553 +0000 UTC m=+5225.657309027" watchObservedRunningTime="2025-11-24 08:14:40.003170515 +0000 UTC m=+5225.659152989" Nov 24 08:14:40 crc kubenswrapper[4799]: I1124 08:14:40.103664 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:40 crc kubenswrapper[4799]: I1124 08:14:40.103738 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:14:41 crc kubenswrapper[4799]: I1124 08:14:41.003770 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7f4a7265-94b0-4631-9500-4ca4e8e4af26","Type":"ContainerStarted","Data":"5a95c14674c6561f0d83ae2ad43fecaae1af5387c51d960792aef78359b491eb"} Nov 24 08:14:41 crc kubenswrapper[4799]: I1124 08:14:41.035981 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.035959445 podStartE2EDuration="8.035959445s" podCreationTimestamp="2025-11-24 08:14:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:14:41.028180724 +0000 UTC m=+5226.684163198" watchObservedRunningTime="2025-11-24 08:14:41.035959445 +0000 UTC m=+5226.691941909" Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.237744 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7nmpk"] Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.240640 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.256746 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7nmpk"] Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.272729 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czdt8\" (UniqueName: \"kubernetes.io/projected/294bb47a-e212-4a04-9c68-e82d3669434a-kube-api-access-czdt8\") pod \"redhat-marketplace-7nmpk\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.272788 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-catalog-content\") pod \"redhat-marketplace-7nmpk\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.272906 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-utilities\") pod \"redhat-marketplace-7nmpk\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.374867 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czdt8\" (UniqueName: \"kubernetes.io/projected/294bb47a-e212-4a04-9c68-e82d3669434a-kube-api-access-czdt8\") pod \"redhat-marketplace-7nmpk\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.374945 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-catalog-content\") pod \"redhat-marketplace-7nmpk\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.374983 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-utilities\") pod \"redhat-marketplace-7nmpk\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.375549 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-catalog-content\") pod \"redhat-marketplace-7nmpk\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.375596 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-utilities\") pod \"redhat-marketplace-7nmpk\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.396185 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czdt8\" (UniqueName: \"kubernetes.io/projected/294bb47a-e212-4a04-9c68-e82d3669434a-kube-api-access-czdt8\") pod \"redhat-marketplace-7nmpk\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:42 crc kubenswrapper[4799]: I1124 08:14:42.608457 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:43 crc kubenswrapper[4799]: I1124 08:14:43.029154 4799 generic.go:334] "Generic (PLEG): container finished" podID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" containerID="ebdb87f7e16966138130773c9eedba1175f6349e5a5085eecb67b7511125f30a" exitCode=0 Nov 24 08:14:43 crc kubenswrapper[4799]: I1124 08:14:43.029432 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nc5j2" event={"ID":"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd","Type":"ContainerDied","Data":"ebdb87f7e16966138130773c9eedba1175f6349e5a5085eecb67b7511125f30a"} Nov 24 08:14:43 crc kubenswrapper[4799]: I1124 08:14:43.115346 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7nmpk"] Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.046334 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nc5j2" event={"ID":"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd","Type":"ContainerStarted","Data":"c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c"} Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.049162 4799 generic.go:334] "Generic (PLEG): container finished" podID="294bb47a-e212-4a04-9c68-e82d3669434a" containerID="dc1254156dc02c80f31596d16f3f588ba9f4372d6e0bc89e5d302511fe7d87e1" exitCode=0 Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.049195 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7nmpk" event={"ID":"294bb47a-e212-4a04-9c68-e82d3669434a","Type":"ContainerDied","Data":"dc1254156dc02c80f31596d16f3f588ba9f4372d6e0bc89e5d302511fe7d87e1"} Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.049210 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7nmpk" event={"ID":"294bb47a-e212-4a04-9c68-e82d3669434a","Type":"ContainerStarted","Data":"fbcd91fc78c4803730fb5be19fc8fd08e45312e5b4dde62857312f417cb9afe6"} Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.082150 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nc5j2" podStartSLOduration=3.328209003 podStartE2EDuration="7.08213358s" podCreationTimestamp="2025-11-24 08:14:37 +0000 UTC" firstStartedPulling="2025-11-24 08:14:39.98817136 +0000 UTC m=+5225.644153834" lastFinishedPulling="2025-11-24 08:14:43.742095937 +0000 UTC m=+5229.398078411" observedRunningTime="2025-11-24 08:14:44.079892146 +0000 UTC m=+5229.735874630" watchObservedRunningTime="2025-11-24 08:14:44.08213358 +0000 UTC m=+5229.738116054" Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.308741 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.308788 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.318190 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.318234 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.347587 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.349399 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.363640 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.369635 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:44 crc kubenswrapper[4799]: I1124 08:14:44.628400 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:14:44 crc kubenswrapper[4799]: E1124 08:14:44.628696 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:14:45 crc kubenswrapper[4799]: I1124 08:14:45.060297 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:45 crc kubenswrapper[4799]: I1124 08:14:45.060342 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 08:14:45 crc kubenswrapper[4799]: I1124 08:14:45.060356 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:45 crc kubenswrapper[4799]: I1124 08:14:45.060366 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 08:14:46 crc kubenswrapper[4799]: I1124 08:14:46.115021 4799 generic.go:334] "Generic (PLEG): container finished" podID="294bb47a-e212-4a04-9c68-e82d3669434a" containerID="bcf2b9bcfef018939fe39deee02acd55a3ecb6be51aa20f01bdb1487b44d30ca" exitCode=0 Nov 24 08:14:46 crc kubenswrapper[4799]: I1124 08:14:46.116966 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7nmpk" event={"ID":"294bb47a-e212-4a04-9c68-e82d3669434a","Type":"ContainerDied","Data":"bcf2b9bcfef018939fe39deee02acd55a3ecb6be51aa20f01bdb1487b44d30ca"} Nov 24 08:14:47 crc kubenswrapper[4799]: I1124 08:14:47.127180 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7nmpk" event={"ID":"294bb47a-e212-4a04-9c68-e82d3669434a","Type":"ContainerStarted","Data":"1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef"} Nov 24 08:14:47 crc kubenswrapper[4799]: I1124 08:14:47.154762 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7nmpk" podStartSLOduration=2.496640687 podStartE2EDuration="5.154730905s" podCreationTimestamp="2025-11-24 08:14:42 +0000 UTC" firstStartedPulling="2025-11-24 08:14:44.050197863 +0000 UTC m=+5229.706180337" lastFinishedPulling="2025-11-24 08:14:46.708288081 +0000 UTC m=+5232.364270555" observedRunningTime="2025-11-24 08:14:47.146264954 +0000 UTC m=+5232.802247429" watchObservedRunningTime="2025-11-24 08:14:47.154730905 +0000 UTC m=+5232.810713429" Nov 24 08:14:47 crc kubenswrapper[4799]: I1124 08:14:47.161817 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:47 crc kubenswrapper[4799]: I1124 08:14:47.162013 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:14:47 crc kubenswrapper[4799]: I1124 08:14:47.242938 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 08:14:47 crc kubenswrapper[4799]: I1124 08:14:47.243075 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:14:47 crc kubenswrapper[4799]: I1124 08:14:47.369738 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 08:14:47 crc kubenswrapper[4799]: I1124 08:14:47.462233 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 08:14:48 crc kubenswrapper[4799]: I1124 08:14:48.208354 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:48 crc kubenswrapper[4799]: I1124 08:14:48.208409 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:48 crc kubenswrapper[4799]: I1124 08:14:48.264559 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:49 crc kubenswrapper[4799]: I1124 08:14:49.203347 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:49 crc kubenswrapper[4799]: I1124 08:14:49.476939 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-57d54fb7f-xkrjn" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.91:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.91:8080: connect: connection refused" Nov 24 08:14:50 crc kubenswrapper[4799]: I1124 08:14:50.105313 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6896b5c567-jgd8d" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.92:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.92:8080: connect: connection refused" Nov 24 08:14:50 crc kubenswrapper[4799]: I1124 08:14:50.417004 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nc5j2"] Nov 24 08:14:51 crc kubenswrapper[4799]: I1124 08:14:51.160456 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nc5j2" podUID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" containerName="registry-server" containerID="cri-o://c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c" gracePeriod=2 Nov 24 08:14:51 crc kubenswrapper[4799]: I1124 08:14:51.665597 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:51 crc kubenswrapper[4799]: I1124 08:14:51.854086 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-utilities\") pod \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " Nov 24 08:14:51 crc kubenswrapper[4799]: I1124 08:14:51.854158 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-catalog-content\") pod \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " Nov 24 08:14:51 crc kubenswrapper[4799]: I1124 08:14:51.854232 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsg5j\" (UniqueName: \"kubernetes.io/projected/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-kube-api-access-jsg5j\") pod \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\" (UID: \"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd\") " Nov 24 08:14:51 crc kubenswrapper[4799]: I1124 08:14:51.857231 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-utilities" (OuterVolumeSpecName: "utilities") pod "bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" (UID: "bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:14:51 crc kubenswrapper[4799]: I1124 08:14:51.872330 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-kube-api-access-jsg5j" (OuterVolumeSpecName: "kube-api-access-jsg5j") pod "bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" (UID: "bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd"). InnerVolumeSpecName "kube-api-access-jsg5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:14:51 crc kubenswrapper[4799]: I1124 08:14:51.913260 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" (UID: "bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:14:51 crc kubenswrapper[4799]: I1124 08:14:51.956355 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:51 crc kubenswrapper[4799]: I1124 08:14:51.956505 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:51 crc kubenswrapper[4799]: I1124 08:14:51.956587 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsg5j\" (UniqueName: \"kubernetes.io/projected/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd-kube-api-access-jsg5j\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.175772 4799 generic.go:334] "Generic (PLEG): container finished" podID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" containerID="c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c" exitCode=0 Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.175939 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nc5j2" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.175978 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nc5j2" event={"ID":"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd","Type":"ContainerDied","Data":"c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c"} Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.176716 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nc5j2" event={"ID":"bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd","Type":"ContainerDied","Data":"da77073b56b1e623cab0b0a66d4e3eadbdb8ee5ffbac3d55666cc190bc177886"} Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.176747 4799 scope.go:117] "RemoveContainer" containerID="c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.221486 4799 scope.go:117] "RemoveContainer" containerID="ebdb87f7e16966138130773c9eedba1175f6349e5a5085eecb67b7511125f30a" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.224958 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nc5j2"] Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.239993 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nc5j2"] Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.270029 4799 scope.go:117] "RemoveContainer" containerID="cec464f06c7c8cabbc0cfe7babefc732b668decafba04d972e1eb5d1258422ee" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.308399 4799 scope.go:117] "RemoveContainer" containerID="c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c" Nov 24 08:14:52 crc kubenswrapper[4799]: E1124 08:14:52.309025 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c\": container with ID starting with c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c not found: ID does not exist" containerID="c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.309070 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c"} err="failed to get container status \"c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c\": rpc error: code = NotFound desc = could not find container \"c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c\": container with ID starting with c7811abb699f68fbe4413ce35536003aed618c5201c600e8d6db7be53d4b1a8c not found: ID does not exist" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.309094 4799 scope.go:117] "RemoveContainer" containerID="ebdb87f7e16966138130773c9eedba1175f6349e5a5085eecb67b7511125f30a" Nov 24 08:14:52 crc kubenswrapper[4799]: E1124 08:14:52.309483 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebdb87f7e16966138130773c9eedba1175f6349e5a5085eecb67b7511125f30a\": container with ID starting with ebdb87f7e16966138130773c9eedba1175f6349e5a5085eecb67b7511125f30a not found: ID does not exist" containerID="ebdb87f7e16966138130773c9eedba1175f6349e5a5085eecb67b7511125f30a" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.309520 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebdb87f7e16966138130773c9eedba1175f6349e5a5085eecb67b7511125f30a"} err="failed to get container status \"ebdb87f7e16966138130773c9eedba1175f6349e5a5085eecb67b7511125f30a\": rpc error: code = NotFound desc = could not find container \"ebdb87f7e16966138130773c9eedba1175f6349e5a5085eecb67b7511125f30a\": container with ID starting with ebdb87f7e16966138130773c9eedba1175f6349e5a5085eecb67b7511125f30a not found: ID does not exist" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.309547 4799 scope.go:117] "RemoveContainer" containerID="cec464f06c7c8cabbc0cfe7babefc732b668decafba04d972e1eb5d1258422ee" Nov 24 08:14:52 crc kubenswrapper[4799]: E1124 08:14:52.309939 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cec464f06c7c8cabbc0cfe7babefc732b668decafba04d972e1eb5d1258422ee\": container with ID starting with cec464f06c7c8cabbc0cfe7babefc732b668decafba04d972e1eb5d1258422ee not found: ID does not exist" containerID="cec464f06c7c8cabbc0cfe7babefc732b668decafba04d972e1eb5d1258422ee" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.309959 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cec464f06c7c8cabbc0cfe7babefc732b668decafba04d972e1eb5d1258422ee"} err="failed to get container status \"cec464f06c7c8cabbc0cfe7babefc732b668decafba04d972e1eb5d1258422ee\": rpc error: code = NotFound desc = could not find container \"cec464f06c7c8cabbc0cfe7babefc732b668decafba04d972e1eb5d1258422ee\": container with ID starting with cec464f06c7c8cabbc0cfe7babefc732b668decafba04d972e1eb5d1258422ee not found: ID does not exist" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.608638 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.608729 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:52 crc kubenswrapper[4799]: I1124 08:14:52.669104 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:53 crc kubenswrapper[4799]: I1124 08:14:53.239176 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:53 crc kubenswrapper[4799]: I1124 08:14:53.639134 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" path="/var/lib/kubelet/pods/bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd/volumes" Nov 24 08:14:54 crc kubenswrapper[4799]: I1124 08:14:54.818750 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7nmpk"] Nov 24 08:14:55 crc kubenswrapper[4799]: I1124 08:14:55.212287 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7nmpk" podUID="294bb47a-e212-4a04-9c68-e82d3669434a" containerName="registry-server" containerID="cri-o://1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef" gracePeriod=2 Nov 24 08:14:55 crc kubenswrapper[4799]: I1124 08:14:55.718802 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:55 crc kubenswrapper[4799]: I1124 08:14:55.760528 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-utilities\") pod \"294bb47a-e212-4a04-9c68-e82d3669434a\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " Nov 24 08:14:55 crc kubenswrapper[4799]: I1124 08:14:55.760625 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czdt8\" (UniqueName: \"kubernetes.io/projected/294bb47a-e212-4a04-9c68-e82d3669434a-kube-api-access-czdt8\") pod \"294bb47a-e212-4a04-9c68-e82d3669434a\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " Nov 24 08:14:55 crc kubenswrapper[4799]: I1124 08:14:55.760785 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-catalog-content\") pod \"294bb47a-e212-4a04-9c68-e82d3669434a\" (UID: \"294bb47a-e212-4a04-9c68-e82d3669434a\") " Nov 24 08:14:55 crc kubenswrapper[4799]: I1124 08:14:55.761502 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-utilities" (OuterVolumeSpecName: "utilities") pod "294bb47a-e212-4a04-9c68-e82d3669434a" (UID: "294bb47a-e212-4a04-9c68-e82d3669434a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:14:55 crc kubenswrapper[4799]: I1124 08:14:55.765930 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/294bb47a-e212-4a04-9c68-e82d3669434a-kube-api-access-czdt8" (OuterVolumeSpecName: "kube-api-access-czdt8") pod "294bb47a-e212-4a04-9c68-e82d3669434a" (UID: "294bb47a-e212-4a04-9c68-e82d3669434a"). InnerVolumeSpecName "kube-api-access-czdt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:14:55 crc kubenswrapper[4799]: I1124 08:14:55.786125 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "294bb47a-e212-4a04-9c68-e82d3669434a" (UID: "294bb47a-e212-4a04-9c68-e82d3669434a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:14:55 crc kubenswrapper[4799]: I1124 08:14:55.862760 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:55 crc kubenswrapper[4799]: I1124 08:14:55.862810 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294bb47a-e212-4a04-9c68-e82d3669434a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:55 crc kubenswrapper[4799]: I1124 08:14:55.862820 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czdt8\" (UniqueName: \"kubernetes.io/projected/294bb47a-e212-4a04-9c68-e82d3669434a-kube-api-access-czdt8\") on node \"crc\" DevicePath \"\"" Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.230296 4799 generic.go:334] "Generic (PLEG): container finished" podID="294bb47a-e212-4a04-9c68-e82d3669434a" containerID="1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef" exitCode=0 Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.230354 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7nmpk" event={"ID":"294bb47a-e212-4a04-9c68-e82d3669434a","Type":"ContainerDied","Data":"1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef"} Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.230394 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7nmpk" Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.230428 4799 scope.go:117] "RemoveContainer" containerID="1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef" Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.230404 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7nmpk" event={"ID":"294bb47a-e212-4a04-9c68-e82d3669434a","Type":"ContainerDied","Data":"fbcd91fc78c4803730fb5be19fc8fd08e45312e5b4dde62857312f417cb9afe6"} Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.273681 4799 scope.go:117] "RemoveContainer" containerID="bcf2b9bcfef018939fe39deee02acd55a3ecb6be51aa20f01bdb1487b44d30ca" Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.298037 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7nmpk"] Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.302734 4799 scope.go:117] "RemoveContainer" containerID="dc1254156dc02c80f31596d16f3f588ba9f4372d6e0bc89e5d302511fe7d87e1" Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.307657 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7nmpk"] Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.344390 4799 scope.go:117] "RemoveContainer" containerID="1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef" Nov 24 08:14:56 crc kubenswrapper[4799]: E1124 08:14:56.344975 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef\": container with ID starting with 1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef not found: ID does not exist" containerID="1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef" Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.345006 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef"} err="failed to get container status \"1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef\": rpc error: code = NotFound desc = could not find container \"1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef\": container with ID starting with 1d38a1651de65bb47b47e150cab5a8d4550cb0c1815a85a9df6d13bcf99177ef not found: ID does not exist" Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.345027 4799 scope.go:117] "RemoveContainer" containerID="bcf2b9bcfef018939fe39deee02acd55a3ecb6be51aa20f01bdb1487b44d30ca" Nov 24 08:14:56 crc kubenswrapper[4799]: E1124 08:14:56.345368 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcf2b9bcfef018939fe39deee02acd55a3ecb6be51aa20f01bdb1487b44d30ca\": container with ID starting with bcf2b9bcfef018939fe39deee02acd55a3ecb6be51aa20f01bdb1487b44d30ca not found: ID does not exist" containerID="bcf2b9bcfef018939fe39deee02acd55a3ecb6be51aa20f01bdb1487b44d30ca" Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.345415 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcf2b9bcfef018939fe39deee02acd55a3ecb6be51aa20f01bdb1487b44d30ca"} err="failed to get container status \"bcf2b9bcfef018939fe39deee02acd55a3ecb6be51aa20f01bdb1487b44d30ca\": rpc error: code = NotFound desc = could not find container \"bcf2b9bcfef018939fe39deee02acd55a3ecb6be51aa20f01bdb1487b44d30ca\": container with ID starting with bcf2b9bcfef018939fe39deee02acd55a3ecb6be51aa20f01bdb1487b44d30ca not found: ID does not exist" Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.345446 4799 scope.go:117] "RemoveContainer" containerID="dc1254156dc02c80f31596d16f3f588ba9f4372d6e0bc89e5d302511fe7d87e1" Nov 24 08:14:56 crc kubenswrapper[4799]: E1124 08:14:56.345738 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc1254156dc02c80f31596d16f3f588ba9f4372d6e0bc89e5d302511fe7d87e1\": container with ID starting with dc1254156dc02c80f31596d16f3f588ba9f4372d6e0bc89e5d302511fe7d87e1 not found: ID does not exist" containerID="dc1254156dc02c80f31596d16f3f588ba9f4372d6e0bc89e5d302511fe7d87e1" Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.345785 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc1254156dc02c80f31596d16f3f588ba9f4372d6e0bc89e5d302511fe7d87e1"} err="failed to get container status \"dc1254156dc02c80f31596d16f3f588ba9f4372d6e0bc89e5d302511fe7d87e1\": rpc error: code = NotFound desc = could not find container \"dc1254156dc02c80f31596d16f3f588ba9f4372d6e0bc89e5d302511fe7d87e1\": container with ID starting with dc1254156dc02c80f31596d16f3f588ba9f4372d6e0bc89e5d302511fe7d87e1 not found: ID does not exist" Nov 24 08:14:56 crc kubenswrapper[4799]: I1124 08:14:56.628746 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:14:56 crc kubenswrapper[4799]: E1124 08:14:56.629000 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:14:57 crc kubenswrapper[4799]: I1124 08:14:57.640270 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="294bb47a-e212-4a04-9c68-e82d3669434a" path="/var/lib/kubelet/pods/294bb47a-e212-4a04-9c68-e82d3669434a/volumes" Nov 24 08:14:59 crc kubenswrapper[4799]: I1124 08:14:59.809825 4799 scope.go:117] "RemoveContainer" containerID="dc14a6488fdb8f0b78a111c5df96c45942e6965d8020be92ae627a6559759dac" Nov 24 08:14:59 crc kubenswrapper[4799]: I1124 08:14:59.840980 4799 scope.go:117] "RemoveContainer" containerID="3b8326afd0a897e5ef200985d1454222061c289ca54efbb6114e721972fbdeed" Nov 24 08:14:59 crc kubenswrapper[4799]: I1124 08:14:59.861770 4799 scope.go:117] "RemoveContainer" containerID="9baa784af9346bce45134500375430672747e10413316507b7bed0357ee4d7d8" Nov 24 08:14:59 crc kubenswrapper[4799]: I1124 08:14:59.879634 4799 scope.go:117] "RemoveContainer" containerID="c1939c2300d3acc7a7b5dcf826e0a316e19046ae2e5527f9484594291f376394" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.148709 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb"] Nov 24 08:15:00 crc kubenswrapper[4799]: E1124 08:15:00.149324 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294bb47a-e212-4a04-9c68-e82d3669434a" containerName="extract-utilities" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.149350 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="294bb47a-e212-4a04-9c68-e82d3669434a" containerName="extract-utilities" Nov 24 08:15:00 crc kubenswrapper[4799]: E1124 08:15:00.149367 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" containerName="extract-utilities" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.149374 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" containerName="extract-utilities" Nov 24 08:15:00 crc kubenswrapper[4799]: E1124 08:15:00.149390 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294bb47a-e212-4a04-9c68-e82d3669434a" containerName="registry-server" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.149398 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="294bb47a-e212-4a04-9c68-e82d3669434a" containerName="registry-server" Nov 24 08:15:00 crc kubenswrapper[4799]: E1124 08:15:00.149419 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" containerName="registry-server" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.149428 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" containerName="registry-server" Nov 24 08:15:00 crc kubenswrapper[4799]: E1124 08:15:00.149437 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" containerName="extract-content" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.149444 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" containerName="extract-content" Nov 24 08:15:00 crc kubenswrapper[4799]: E1124 08:15:00.149466 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294bb47a-e212-4a04-9c68-e82d3669434a" containerName="extract-content" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.149476 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="294bb47a-e212-4a04-9c68-e82d3669434a" containerName="extract-content" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.149637 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="294bb47a-e212-4a04-9c68-e82d3669434a" containerName="registry-server" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.149660 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbf0c9d8-5382-4ccc-b433-c63cd5cf7acd" containerName="registry-server" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.150384 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.158470 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.158994 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.176797 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb"] Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.249534 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2836c1d2-0310-411e-8895-c294f6a09c89-config-volume\") pod \"collect-profiles-29399535-sd8cb\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.249960 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsrm4\" (UniqueName: \"kubernetes.io/projected/2836c1d2-0310-411e-8895-c294f6a09c89-kube-api-access-hsrm4\") pod \"collect-profiles-29399535-sd8cb\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.250147 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2836c1d2-0310-411e-8895-c294f6a09c89-secret-volume\") pod \"collect-profiles-29399535-sd8cb\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.351884 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2836c1d2-0310-411e-8895-c294f6a09c89-config-volume\") pod \"collect-profiles-29399535-sd8cb\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.352056 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsrm4\" (UniqueName: \"kubernetes.io/projected/2836c1d2-0310-411e-8895-c294f6a09c89-kube-api-access-hsrm4\") pod \"collect-profiles-29399535-sd8cb\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.352267 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2836c1d2-0310-411e-8895-c294f6a09c89-secret-volume\") pod \"collect-profiles-29399535-sd8cb\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.354546 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2836c1d2-0310-411e-8895-c294f6a09c89-config-volume\") pod \"collect-profiles-29399535-sd8cb\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.361420 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2836c1d2-0310-411e-8895-c294f6a09c89-secret-volume\") pod \"collect-profiles-29399535-sd8cb\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.376517 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsrm4\" (UniqueName: \"kubernetes.io/projected/2836c1d2-0310-411e-8895-c294f6a09c89-kube-api-access-hsrm4\") pod \"collect-profiles-29399535-sd8cb\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.499177 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:00 crc kubenswrapper[4799]: W1124 08:15:00.974988 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2836c1d2_0310_411e_8895_c294f6a09c89.slice/crio-f176d3e51305a90408ad470d03b93fb4100e7b2ef8495eb347de495c204e8af5 WatchSource:0}: Error finding container f176d3e51305a90408ad470d03b93fb4100e7b2ef8495eb347de495c204e8af5: Status 404 returned error can't find the container with id f176d3e51305a90408ad470d03b93fb4100e7b2ef8495eb347de495c204e8af5 Nov 24 08:15:00 crc kubenswrapper[4799]: I1124 08:15:00.980761 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb"] Nov 24 08:15:01 crc kubenswrapper[4799]: I1124 08:15:01.279173 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" event={"ID":"2836c1d2-0310-411e-8895-c294f6a09c89","Type":"ContainerStarted","Data":"f47d0bea02c82e54e505a7f5e5169d08633bc41c6f3cf960608787e9054103b0"} Nov 24 08:15:01 crc kubenswrapper[4799]: I1124 08:15:01.279219 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" event={"ID":"2836c1d2-0310-411e-8895-c294f6a09c89","Type":"ContainerStarted","Data":"f176d3e51305a90408ad470d03b93fb4100e7b2ef8495eb347de495c204e8af5"} Nov 24 08:15:01 crc kubenswrapper[4799]: I1124 08:15:01.300060 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" podStartSLOduration=1.300042685 podStartE2EDuration="1.300042685s" podCreationTimestamp="2025-11-24 08:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:15:01.299343185 +0000 UTC m=+5246.955325659" watchObservedRunningTime="2025-11-24 08:15:01.300042685 +0000 UTC m=+5246.956025159" Nov 24 08:15:01 crc kubenswrapper[4799]: I1124 08:15:01.307683 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:15:01 crc kubenswrapper[4799]: I1124 08:15:01.937836 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:15:02 crc kubenswrapper[4799]: I1124 08:15:02.302526 4799 generic.go:334] "Generic (PLEG): container finished" podID="2836c1d2-0310-411e-8895-c294f6a09c89" containerID="f47d0bea02c82e54e505a7f5e5169d08633bc41c6f3cf960608787e9054103b0" exitCode=0 Nov 24 08:15:02 crc kubenswrapper[4799]: I1124 08:15:02.302629 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" event={"ID":"2836c1d2-0310-411e-8895-c294f6a09c89","Type":"ContainerDied","Data":"f47d0bea02c82e54e505a7f5e5169d08633bc41c6f3cf960608787e9054103b0"} Nov 24 08:15:02 crc kubenswrapper[4799]: I1124 08:15:02.976587 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.622058 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.627367 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.693342 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-57d54fb7f-xkrjn"] Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.694154 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-57d54fb7f-xkrjn" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon-log" containerID="cri-o://3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c" gracePeriod=30 Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.694402 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-57d54fb7f-xkrjn" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon" containerID="cri-o://a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd" gracePeriod=30 Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.722637 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2836c1d2-0310-411e-8895-c294f6a09c89-secret-volume\") pod \"2836c1d2-0310-411e-8895-c294f6a09c89\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.723020 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsrm4\" (UniqueName: \"kubernetes.io/projected/2836c1d2-0310-411e-8895-c294f6a09c89-kube-api-access-hsrm4\") pod \"2836c1d2-0310-411e-8895-c294f6a09c89\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.723197 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2836c1d2-0310-411e-8895-c294f6a09c89-config-volume\") pod \"2836c1d2-0310-411e-8895-c294f6a09c89\" (UID: \"2836c1d2-0310-411e-8895-c294f6a09c89\") " Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.726430 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2836c1d2-0310-411e-8895-c294f6a09c89-config-volume" (OuterVolumeSpecName: "config-volume") pod "2836c1d2-0310-411e-8895-c294f6a09c89" (UID: "2836c1d2-0310-411e-8895-c294f6a09c89"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.733453 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2836c1d2-0310-411e-8895-c294f6a09c89-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2836c1d2-0310-411e-8895-c294f6a09c89" (UID: "2836c1d2-0310-411e-8895-c294f6a09c89"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.734156 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2836c1d2-0310-411e-8895-c294f6a09c89-kube-api-access-hsrm4" (OuterVolumeSpecName: "kube-api-access-hsrm4") pod "2836c1d2-0310-411e-8895-c294f6a09c89" (UID: "2836c1d2-0310-411e-8895-c294f6a09c89"). InnerVolumeSpecName "kube-api-access-hsrm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.826827 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2836c1d2-0310-411e-8895-c294f6a09c89-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.826893 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsrm4\" (UniqueName: \"kubernetes.io/projected/2836c1d2-0310-411e-8895-c294f6a09c89-kube-api-access-hsrm4\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:03 crc kubenswrapper[4799]: I1124 08:15:03.826908 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2836c1d2-0310-411e-8895-c294f6a09c89-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:04 crc kubenswrapper[4799]: I1124 08:15:04.324541 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" event={"ID":"2836c1d2-0310-411e-8895-c294f6a09c89","Type":"ContainerDied","Data":"f176d3e51305a90408ad470d03b93fb4100e7b2ef8495eb347de495c204e8af5"} Nov 24 08:15:04 crc kubenswrapper[4799]: I1124 08:15:04.325008 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f176d3e51305a90408ad470d03b93fb4100e7b2ef8495eb347de495c204e8af5" Nov 24 08:15:04 crc kubenswrapper[4799]: I1124 08:15:04.324653 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb" Nov 24 08:15:04 crc kubenswrapper[4799]: I1124 08:15:04.396777 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8"] Nov 24 08:15:04 crc kubenswrapper[4799]: I1124 08:15:04.406875 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399490-7wgp8"] Nov 24 08:15:05 crc kubenswrapper[4799]: I1124 08:15:05.644782 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a8f2fb6-f858-45ee-b75e-9d2e5f55578b" path="/var/lib/kubelet/pods/6a8f2fb6-f858-45ee-b75e-9d2e5f55578b/volumes" Nov 24 08:15:07 crc kubenswrapper[4799]: I1124 08:15:07.358777 4799 generic.go:334] "Generic (PLEG): container finished" podID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerID="a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd" exitCode=0 Nov 24 08:15:07 crc kubenswrapper[4799]: I1124 08:15:07.358887 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d54fb7f-xkrjn" event={"ID":"2e498b0a-27e6-4e14-9440-48e18e327e70","Type":"ContainerDied","Data":"a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd"} Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.376530 4799 generic.go:334] "Generic (PLEG): container finished" podID="b85335a2-d197-47b5-be68-e45201d3657f" containerID="d009d073bd9d809d5a3bf3a99ec77b9e48ca66e7bc1fcf6cf11932881e2b264e" exitCode=137 Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.376767 4799 generic.go:334] "Generic (PLEG): container finished" podID="b85335a2-d197-47b5-be68-e45201d3657f" containerID="7eff771f603f594f1c891f04bffed0b19dce8367713050adf97f181b9a2ecf49" exitCode=137 Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.376786 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bccb8fd89-sll6f" event={"ID":"b85335a2-d197-47b5-be68-e45201d3657f","Type":"ContainerDied","Data":"d009d073bd9d809d5a3bf3a99ec77b9e48ca66e7bc1fcf6cf11932881e2b264e"} Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.376810 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bccb8fd89-sll6f" event={"ID":"b85335a2-d197-47b5-be68-e45201d3657f","Type":"ContainerDied","Data":"7eff771f603f594f1c891f04bffed0b19dce8367713050adf97f181b9a2ecf49"} Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.376819 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bccb8fd89-sll6f" event={"ID":"b85335a2-d197-47b5-be68-e45201d3657f","Type":"ContainerDied","Data":"33b44145a34e2a77449221b5fb927d7043b856d2fe49a74291ce962b193c68d7"} Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.376829 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33b44145a34e2a77449221b5fb927d7043b856d2fe49a74291ce962b193c68d7" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.383949 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.474454 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-57d54fb7f-xkrjn" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.91:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.91:8080: connect: connection refused" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.519496 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tlq4\" (UniqueName: \"kubernetes.io/projected/b85335a2-d197-47b5-be68-e45201d3657f-kube-api-access-9tlq4\") pod \"b85335a2-d197-47b5-be68-e45201d3657f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.519617 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-config-data\") pod \"b85335a2-d197-47b5-be68-e45201d3657f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.519698 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-scripts\") pod \"b85335a2-d197-47b5-be68-e45201d3657f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.519735 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b85335a2-d197-47b5-be68-e45201d3657f-logs\") pod \"b85335a2-d197-47b5-be68-e45201d3657f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.519759 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b85335a2-d197-47b5-be68-e45201d3657f-horizon-secret-key\") pod \"b85335a2-d197-47b5-be68-e45201d3657f\" (UID: \"b85335a2-d197-47b5-be68-e45201d3657f\") " Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.520859 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b85335a2-d197-47b5-be68-e45201d3657f-logs" (OuterVolumeSpecName: "logs") pod "b85335a2-d197-47b5-be68-e45201d3657f" (UID: "b85335a2-d197-47b5-be68-e45201d3657f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.525240 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85335a2-d197-47b5-be68-e45201d3657f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b85335a2-d197-47b5-be68-e45201d3657f" (UID: "b85335a2-d197-47b5-be68-e45201d3657f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.525611 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b85335a2-d197-47b5-be68-e45201d3657f-kube-api-access-9tlq4" (OuterVolumeSpecName: "kube-api-access-9tlq4") pod "b85335a2-d197-47b5-be68-e45201d3657f" (UID: "b85335a2-d197-47b5-be68-e45201d3657f"). InnerVolumeSpecName "kube-api-access-9tlq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.542751 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-config-data" (OuterVolumeSpecName: "config-data") pod "b85335a2-d197-47b5-be68-e45201d3657f" (UID: "b85335a2-d197-47b5-be68-e45201d3657f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.555084 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-scripts" (OuterVolumeSpecName: "scripts") pod "b85335a2-d197-47b5-be68-e45201d3657f" (UID: "b85335a2-d197-47b5-be68-e45201d3657f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.621517 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tlq4\" (UniqueName: \"kubernetes.io/projected/b85335a2-d197-47b5-be68-e45201d3657f-kube-api-access-9tlq4\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.621549 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.621560 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b85335a2-d197-47b5-be68-e45201d3657f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.621571 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b85335a2-d197-47b5-be68-e45201d3657f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:09 crc kubenswrapper[4799]: I1124 08:15:09.621600 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b85335a2-d197-47b5-be68-e45201d3657f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:10 crc kubenswrapper[4799]: I1124 08:15:10.385932 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bccb8fd89-sll6f" Nov 24 08:15:10 crc kubenswrapper[4799]: I1124 08:15:10.413729 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bccb8fd89-sll6f"] Nov 24 08:15:10 crc kubenswrapper[4799]: I1124 08:15:10.423427 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5bccb8fd89-sll6f"] Nov 24 08:15:10 crc kubenswrapper[4799]: I1124 08:15:10.628357 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:15:10 crc kubenswrapper[4799]: E1124 08:15:10.628809 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:15:11 crc kubenswrapper[4799]: I1124 08:15:11.639770 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b85335a2-d197-47b5-be68-e45201d3657f" path="/var/lib/kubelet/pods/b85335a2-d197-47b5-be68-e45201d3657f/volumes" Nov 24 08:15:19 crc kubenswrapper[4799]: I1124 08:15:19.475428 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-57d54fb7f-xkrjn" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.91:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.91:8080: connect: connection refused" Nov 24 08:15:25 crc kubenswrapper[4799]: I1124 08:15:25.641354 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:15:26 crc kubenswrapper[4799]: I1124 08:15:26.609371 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"7ac03e80c7a163ff61b627bd7ec43f0aeff1db92dce99a8b10b286b57ba4f7a4"} Nov 24 08:15:29 crc kubenswrapper[4799]: I1124 08:15:29.476409 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-57d54fb7f-xkrjn" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.91:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.91:8080: connect: connection refused" Nov 24 08:15:29 crc kubenswrapper[4799]: I1124 08:15:29.477351 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.120289 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.236770 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-scripts\") pod \"2e498b0a-27e6-4e14-9440-48e18e327e70\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.236839 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-config-data\") pod \"2e498b0a-27e6-4e14-9440-48e18e327e70\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.236925 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e498b0a-27e6-4e14-9440-48e18e327e70-horizon-secret-key\") pod \"2e498b0a-27e6-4e14-9440-48e18e327e70\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.236990 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwdbm\" (UniqueName: \"kubernetes.io/projected/2e498b0a-27e6-4e14-9440-48e18e327e70-kube-api-access-xwdbm\") pod \"2e498b0a-27e6-4e14-9440-48e18e327e70\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.237050 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e498b0a-27e6-4e14-9440-48e18e327e70-logs\") pod \"2e498b0a-27e6-4e14-9440-48e18e327e70\" (UID: \"2e498b0a-27e6-4e14-9440-48e18e327e70\") " Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.237721 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e498b0a-27e6-4e14-9440-48e18e327e70-logs" (OuterVolumeSpecName: "logs") pod "2e498b0a-27e6-4e14-9440-48e18e327e70" (UID: "2e498b0a-27e6-4e14-9440-48e18e327e70"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.242466 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e498b0a-27e6-4e14-9440-48e18e327e70-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2e498b0a-27e6-4e14-9440-48e18e327e70" (UID: "2e498b0a-27e6-4e14-9440-48e18e327e70"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.242842 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e498b0a-27e6-4e14-9440-48e18e327e70-kube-api-access-xwdbm" (OuterVolumeSpecName: "kube-api-access-xwdbm") pod "2e498b0a-27e6-4e14-9440-48e18e327e70" (UID: "2e498b0a-27e6-4e14-9440-48e18e327e70"). InnerVolumeSpecName "kube-api-access-xwdbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.260146 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-config-data" (OuterVolumeSpecName: "config-data") pod "2e498b0a-27e6-4e14-9440-48e18e327e70" (UID: "2e498b0a-27e6-4e14-9440-48e18e327e70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.265904 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-scripts" (OuterVolumeSpecName: "scripts") pod "2e498b0a-27e6-4e14-9440-48e18e327e70" (UID: "2e498b0a-27e6-4e14-9440-48e18e327e70"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.339096 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e498b0a-27e6-4e14-9440-48e18e327e70-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.339543 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwdbm\" (UniqueName: \"kubernetes.io/projected/2e498b0a-27e6-4e14-9440-48e18e327e70-kube-api-access-xwdbm\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.339613 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e498b0a-27e6-4e14-9440-48e18e327e70-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.339670 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.339731 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e498b0a-27e6-4e14-9440-48e18e327e70-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.700919 4799 generic.go:334] "Generic (PLEG): container finished" podID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerID="3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c" exitCode=137 Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.700980 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57d54fb7f-xkrjn" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.701029 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d54fb7f-xkrjn" event={"ID":"2e498b0a-27e6-4e14-9440-48e18e327e70","Type":"ContainerDied","Data":"3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c"} Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.701911 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d54fb7f-xkrjn" event={"ID":"2e498b0a-27e6-4e14-9440-48e18e327e70","Type":"ContainerDied","Data":"604f80a945ffa378752cd30a785981143298632516d586f3549a32cb58764cc3"} Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.701963 4799 scope.go:117] "RemoveContainer" containerID="a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.756677 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-57d54fb7f-xkrjn"] Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.767285 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-57d54fb7f-xkrjn"] Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.924466 4799 scope.go:117] "RemoveContainer" containerID="3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.943187 4799 scope.go:117] "RemoveContainer" containerID="a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd" Nov 24 08:15:34 crc kubenswrapper[4799]: E1124 08:15:34.943810 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd\": container with ID starting with a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd not found: ID does not exist" containerID="a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.943899 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd"} err="failed to get container status \"a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd\": rpc error: code = NotFound desc = could not find container \"a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd\": container with ID starting with a3529057a779f7e2d3ca659df12de8a8e0657330a8fadc5f23d60f5cf7fe52dd not found: ID does not exist" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.943931 4799 scope.go:117] "RemoveContainer" containerID="3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c" Nov 24 08:15:34 crc kubenswrapper[4799]: E1124 08:15:34.944363 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c\": container with ID starting with 3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c not found: ID does not exist" containerID="3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c" Nov 24 08:15:34 crc kubenswrapper[4799]: I1124 08:15:34.944462 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c"} err="failed to get container status \"3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c\": rpc error: code = NotFound desc = could not find container \"3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c\": container with ID starting with 3066c61f8777e5e98b3be9c01e3cfd9e1b5086e0f58c29b2c9a78d4bb8af8c7c not found: ID does not exist" Nov 24 08:15:35 crc kubenswrapper[4799]: I1124 08:15:35.650218 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" path="/var/lib/kubelet/pods/2e498b0a-27e6-4e14-9440-48e18e327e70/volumes" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.644358 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7cd9788b67-pld4l"] Nov 24 08:15:46 crc kubenswrapper[4799]: E1124 08:15:46.645306 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85335a2-d197-47b5-be68-e45201d3657f" containerName="horizon-log" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.645323 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85335a2-d197-47b5-be68-e45201d3657f" containerName="horizon-log" Nov 24 08:15:46 crc kubenswrapper[4799]: E1124 08:15:46.645368 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2836c1d2-0310-411e-8895-c294f6a09c89" containerName="collect-profiles" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.645380 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2836c1d2-0310-411e-8895-c294f6a09c89" containerName="collect-profiles" Nov 24 08:15:46 crc kubenswrapper[4799]: E1124 08:15:46.645407 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.645414 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon" Nov 24 08:15:46 crc kubenswrapper[4799]: E1124 08:15:46.645433 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85335a2-d197-47b5-be68-e45201d3657f" containerName="horizon" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.645451 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85335a2-d197-47b5-be68-e45201d3657f" containerName="horizon" Nov 24 08:15:46 crc kubenswrapper[4799]: E1124 08:15:46.645468 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon-log" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.645475 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon-log" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.645676 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b85335a2-d197-47b5-be68-e45201d3657f" containerName="horizon-log" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.645695 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b85335a2-d197-47b5-be68-e45201d3657f" containerName="horizon" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.645713 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon-log" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.645725 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e498b0a-27e6-4e14-9440-48e18e327e70" containerName="horizon" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.645741 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2836c1d2-0310-411e-8895-c294f6a09c89" containerName="collect-profiles" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.646940 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.656965 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cd9788b67-pld4l"] Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.823730 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75ed23b8-89f0-4527-bff2-5c8e60757c8f-horizon-secret-key\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.824552 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75ed23b8-89f0-4527-bff2-5c8e60757c8f-scripts\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.824668 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75ed23b8-89f0-4527-bff2-5c8e60757c8f-config-data\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.824737 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75ed23b8-89f0-4527-bff2-5c8e60757c8f-logs\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.824804 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhr9n\" (UniqueName: \"kubernetes.io/projected/75ed23b8-89f0-4527-bff2-5c8e60757c8f-kube-api-access-vhr9n\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.926226 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75ed23b8-89f0-4527-bff2-5c8e60757c8f-horizon-secret-key\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.926327 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75ed23b8-89f0-4527-bff2-5c8e60757c8f-scripts\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.926379 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75ed23b8-89f0-4527-bff2-5c8e60757c8f-config-data\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.926399 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75ed23b8-89f0-4527-bff2-5c8e60757c8f-logs\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.926426 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhr9n\" (UniqueName: \"kubernetes.io/projected/75ed23b8-89f0-4527-bff2-5c8e60757c8f-kube-api-access-vhr9n\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.928164 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75ed23b8-89f0-4527-bff2-5c8e60757c8f-scripts\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.928164 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75ed23b8-89f0-4527-bff2-5c8e60757c8f-logs\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.929161 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75ed23b8-89f0-4527-bff2-5c8e60757c8f-config-data\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.947157 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhr9n\" (UniqueName: \"kubernetes.io/projected/75ed23b8-89f0-4527-bff2-5c8e60757c8f-kube-api-access-vhr9n\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:46 crc kubenswrapper[4799]: I1124 08:15:46.947506 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75ed23b8-89f0-4527-bff2-5c8e60757c8f-horizon-secret-key\") pod \"horizon-7cd9788b67-pld4l\" (UID: \"75ed23b8-89f0-4527-bff2-5c8e60757c8f\") " pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.021349 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.516410 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cd9788b67-pld4l"] Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.748038 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-pzp8c"] Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.749666 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pzp8c" Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.758659 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pzp8c"] Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.843484 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cd9788b67-pld4l" event={"ID":"75ed23b8-89f0-4527-bff2-5c8e60757c8f","Type":"ContainerStarted","Data":"ef34ba7eb34c73e27015fa24a00a72c54ab9141e31ad45feb248cdd1cec5893a"} Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.843672 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cd9788b67-pld4l" event={"ID":"75ed23b8-89f0-4527-bff2-5c8e60757c8f","Type":"ContainerStarted","Data":"56392ea2457a5adb14aa75830cb6cf2220da4652f0c9bac1816c52011e5cf833"} Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.855290 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-7167-account-create-nxvxk"] Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.857263 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-7167-account-create-nxvxk" Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.861939 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.868566 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-7167-account-create-nxvxk"] Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.945161 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlprs\" (UniqueName: \"kubernetes.io/projected/5bb0216b-56a7-4822-bfce-0cf463a9d595-kube-api-access-wlprs\") pod \"heat-db-create-pzp8c\" (UID: \"5bb0216b-56a7-4822-bfce-0cf463a9d595\") " pod="openstack/heat-db-create-pzp8c" Nov 24 08:15:47 crc kubenswrapper[4799]: I1124 08:15:47.945774 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bb0216b-56a7-4822-bfce-0cf463a9d595-operator-scripts\") pod \"heat-db-create-pzp8c\" (UID: \"5bb0216b-56a7-4822-bfce-0cf463a9d595\") " pod="openstack/heat-db-create-pzp8c" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.047753 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bb0216b-56a7-4822-bfce-0cf463a9d595-operator-scripts\") pod \"heat-db-create-pzp8c\" (UID: \"5bb0216b-56a7-4822-bfce-0cf463a9d595\") " pod="openstack/heat-db-create-pzp8c" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.047866 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klv2w\" (UniqueName: \"kubernetes.io/projected/388f8341-6e2d-40e3-b35c-d1bd0d95e134-kube-api-access-klv2w\") pod \"heat-7167-account-create-nxvxk\" (UID: \"388f8341-6e2d-40e3-b35c-d1bd0d95e134\") " pod="openstack/heat-7167-account-create-nxvxk" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.047894 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/388f8341-6e2d-40e3-b35c-d1bd0d95e134-operator-scripts\") pod \"heat-7167-account-create-nxvxk\" (UID: \"388f8341-6e2d-40e3-b35c-d1bd0d95e134\") " pod="openstack/heat-7167-account-create-nxvxk" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.048056 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlprs\" (UniqueName: \"kubernetes.io/projected/5bb0216b-56a7-4822-bfce-0cf463a9d595-kube-api-access-wlprs\") pod \"heat-db-create-pzp8c\" (UID: \"5bb0216b-56a7-4822-bfce-0cf463a9d595\") " pod="openstack/heat-db-create-pzp8c" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.048563 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bb0216b-56a7-4822-bfce-0cf463a9d595-operator-scripts\") pod \"heat-db-create-pzp8c\" (UID: \"5bb0216b-56a7-4822-bfce-0cf463a9d595\") " pod="openstack/heat-db-create-pzp8c" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.065144 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlprs\" (UniqueName: \"kubernetes.io/projected/5bb0216b-56a7-4822-bfce-0cf463a9d595-kube-api-access-wlprs\") pod \"heat-db-create-pzp8c\" (UID: \"5bb0216b-56a7-4822-bfce-0cf463a9d595\") " pod="openstack/heat-db-create-pzp8c" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.119386 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pzp8c" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.149731 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klv2w\" (UniqueName: \"kubernetes.io/projected/388f8341-6e2d-40e3-b35c-d1bd0d95e134-kube-api-access-klv2w\") pod \"heat-7167-account-create-nxvxk\" (UID: \"388f8341-6e2d-40e3-b35c-d1bd0d95e134\") " pod="openstack/heat-7167-account-create-nxvxk" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.149778 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/388f8341-6e2d-40e3-b35c-d1bd0d95e134-operator-scripts\") pod \"heat-7167-account-create-nxvxk\" (UID: \"388f8341-6e2d-40e3-b35c-d1bd0d95e134\") " pod="openstack/heat-7167-account-create-nxvxk" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.150643 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/388f8341-6e2d-40e3-b35c-d1bd0d95e134-operator-scripts\") pod \"heat-7167-account-create-nxvxk\" (UID: \"388f8341-6e2d-40e3-b35c-d1bd0d95e134\") " pod="openstack/heat-7167-account-create-nxvxk" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.172768 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klv2w\" (UniqueName: \"kubernetes.io/projected/388f8341-6e2d-40e3-b35c-d1bd0d95e134-kube-api-access-klv2w\") pod \"heat-7167-account-create-nxvxk\" (UID: \"388f8341-6e2d-40e3-b35c-d1bd0d95e134\") " pod="openstack/heat-7167-account-create-nxvxk" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.192495 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-7167-account-create-nxvxk" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.666919 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pzp8c"] Nov 24 08:15:48 crc kubenswrapper[4799]: W1124 08:15:48.671733 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bb0216b_56a7_4822_bfce_0cf463a9d595.slice/crio-3d896e46aa72208ce22cd85060b920b2e478af2d6fc3fa3c6b898a751555f494 WatchSource:0}: Error finding container 3d896e46aa72208ce22cd85060b920b2e478af2d6fc3fa3c6b898a751555f494: Status 404 returned error can't find the container with id 3d896e46aa72208ce22cd85060b920b2e478af2d6fc3fa3c6b898a751555f494 Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.771462 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-7167-account-create-nxvxk"] Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.852948 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-7167-account-create-nxvxk" event={"ID":"388f8341-6e2d-40e3-b35c-d1bd0d95e134","Type":"ContainerStarted","Data":"5110d7d0767ed68f6efb08255fdbfefcf85c8ab788c403e337cdcb4d9fd90b08"} Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.854811 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cd9788b67-pld4l" event={"ID":"75ed23b8-89f0-4527-bff2-5c8e60757c8f","Type":"ContainerStarted","Data":"a0c96e8e9d069f1f8bee3b5516ea6f7f0ff5eb8a70dbf61ffcaf24eb20c1d71b"} Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.857381 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pzp8c" event={"ID":"5bb0216b-56a7-4822-bfce-0cf463a9d595","Type":"ContainerStarted","Data":"f3b03d3bdd29c910243b0c4ab8146de727a5be380550f3c756f2cf63400c6576"} Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.857421 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pzp8c" event={"ID":"5bb0216b-56a7-4822-bfce-0cf463a9d595","Type":"ContainerStarted","Data":"3d896e46aa72208ce22cd85060b920b2e478af2d6fc3fa3c6b898a751555f494"} Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.882652 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7cd9788b67-pld4l" podStartSLOduration=2.882619113 podStartE2EDuration="2.882619113s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:15:48.873663208 +0000 UTC m=+5294.529645672" watchObservedRunningTime="2025-11-24 08:15:48.882619113 +0000 UTC m=+5294.538601597" Nov 24 08:15:48 crc kubenswrapper[4799]: I1124 08:15:48.926807 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-pzp8c" podStartSLOduration=1.926787067 podStartE2EDuration="1.926787067s" podCreationTimestamp="2025-11-24 08:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:15:48.926065147 +0000 UTC m=+5294.582047611" watchObservedRunningTime="2025-11-24 08:15:48.926787067 +0000 UTC m=+5294.582769541" Nov 24 08:15:49 crc kubenswrapper[4799]: I1124 08:15:49.878122 4799 generic.go:334] "Generic (PLEG): container finished" podID="388f8341-6e2d-40e3-b35c-d1bd0d95e134" containerID="d3abb67370ec878d7c0d6f808292579923792f83eb0f919b804aac968f4ee28a" exitCode=0 Nov 24 08:15:49 crc kubenswrapper[4799]: I1124 08:15:49.878532 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-7167-account-create-nxvxk" event={"ID":"388f8341-6e2d-40e3-b35c-d1bd0d95e134","Type":"ContainerDied","Data":"d3abb67370ec878d7c0d6f808292579923792f83eb0f919b804aac968f4ee28a"} Nov 24 08:15:49 crc kubenswrapper[4799]: I1124 08:15:49.884688 4799 generic.go:334] "Generic (PLEG): container finished" podID="5bb0216b-56a7-4822-bfce-0cf463a9d595" containerID="f3b03d3bdd29c910243b0c4ab8146de727a5be380550f3c756f2cf63400c6576" exitCode=0 Nov 24 08:15:49 crc kubenswrapper[4799]: I1124 08:15:49.885581 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pzp8c" event={"ID":"5bb0216b-56a7-4822-bfce-0cf463a9d595","Type":"ContainerDied","Data":"f3b03d3bdd29c910243b0c4ab8146de727a5be380550f3c756f2cf63400c6576"} Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.318914 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pzp8c" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.326483 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-7167-account-create-nxvxk" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.513066 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlprs\" (UniqueName: \"kubernetes.io/projected/5bb0216b-56a7-4822-bfce-0cf463a9d595-kube-api-access-wlprs\") pod \"5bb0216b-56a7-4822-bfce-0cf463a9d595\" (UID: \"5bb0216b-56a7-4822-bfce-0cf463a9d595\") " Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.513221 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/388f8341-6e2d-40e3-b35c-d1bd0d95e134-operator-scripts\") pod \"388f8341-6e2d-40e3-b35c-d1bd0d95e134\" (UID: \"388f8341-6e2d-40e3-b35c-d1bd0d95e134\") " Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.513261 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bb0216b-56a7-4822-bfce-0cf463a9d595-operator-scripts\") pod \"5bb0216b-56a7-4822-bfce-0cf463a9d595\" (UID: \"5bb0216b-56a7-4822-bfce-0cf463a9d595\") " Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.513312 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klv2w\" (UniqueName: \"kubernetes.io/projected/388f8341-6e2d-40e3-b35c-d1bd0d95e134-kube-api-access-klv2w\") pod \"388f8341-6e2d-40e3-b35c-d1bd0d95e134\" (UID: \"388f8341-6e2d-40e3-b35c-d1bd0d95e134\") " Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.513726 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/388f8341-6e2d-40e3-b35c-d1bd0d95e134-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "388f8341-6e2d-40e3-b35c-d1bd0d95e134" (UID: "388f8341-6e2d-40e3-b35c-d1bd0d95e134"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.513819 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bb0216b-56a7-4822-bfce-0cf463a9d595-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5bb0216b-56a7-4822-bfce-0cf463a9d595" (UID: "5bb0216b-56a7-4822-bfce-0cf463a9d595"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.519034 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/388f8341-6e2d-40e3-b35c-d1bd0d95e134-kube-api-access-klv2w" (OuterVolumeSpecName: "kube-api-access-klv2w") pod "388f8341-6e2d-40e3-b35c-d1bd0d95e134" (UID: "388f8341-6e2d-40e3-b35c-d1bd0d95e134"). InnerVolumeSpecName "kube-api-access-klv2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.534026 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bb0216b-56a7-4822-bfce-0cf463a9d595-kube-api-access-wlprs" (OuterVolumeSpecName: "kube-api-access-wlprs") pod "5bb0216b-56a7-4822-bfce-0cf463a9d595" (UID: "5bb0216b-56a7-4822-bfce-0cf463a9d595"). InnerVolumeSpecName "kube-api-access-wlprs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.616409 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlprs\" (UniqueName: \"kubernetes.io/projected/5bb0216b-56a7-4822-bfce-0cf463a9d595-kube-api-access-wlprs\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.616476 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/388f8341-6e2d-40e3-b35c-d1bd0d95e134-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.616497 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bb0216b-56a7-4822-bfce-0cf463a9d595-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.616516 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klv2w\" (UniqueName: \"kubernetes.io/projected/388f8341-6e2d-40e3-b35c-d1bd0d95e134-kube-api-access-klv2w\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.908563 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-7167-account-create-nxvxk" event={"ID":"388f8341-6e2d-40e3-b35c-d1bd0d95e134","Type":"ContainerDied","Data":"5110d7d0767ed68f6efb08255fdbfefcf85c8ab788c403e337cdcb4d9fd90b08"} Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.908752 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5110d7d0767ed68f6efb08255fdbfefcf85c8ab788c403e337cdcb4d9fd90b08" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.908603 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-7167-account-create-nxvxk" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.911066 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pzp8c" Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.911036 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pzp8c" event={"ID":"5bb0216b-56a7-4822-bfce-0cf463a9d595","Type":"ContainerDied","Data":"3d896e46aa72208ce22cd85060b920b2e478af2d6fc3fa3c6b898a751555f494"} Nov 24 08:15:51 crc kubenswrapper[4799]: I1124 08:15:51.912030 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d896e46aa72208ce22cd85060b920b2e478af2d6fc3fa3c6b898a751555f494" Nov 24 08:15:52 crc kubenswrapper[4799]: I1124 08:15:52.989613 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-7bnw6"] Nov 24 08:15:52 crc kubenswrapper[4799]: E1124 08:15:52.992133 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="388f8341-6e2d-40e3-b35c-d1bd0d95e134" containerName="mariadb-account-create" Nov 24 08:15:52 crc kubenswrapper[4799]: I1124 08:15:52.992327 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="388f8341-6e2d-40e3-b35c-d1bd0d95e134" containerName="mariadb-account-create" Nov 24 08:15:52 crc kubenswrapper[4799]: E1124 08:15:52.992513 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bb0216b-56a7-4822-bfce-0cf463a9d595" containerName="mariadb-database-create" Nov 24 08:15:52 crc kubenswrapper[4799]: I1124 08:15:52.992642 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bb0216b-56a7-4822-bfce-0cf463a9d595" containerName="mariadb-database-create" Nov 24 08:15:52 crc kubenswrapper[4799]: I1124 08:15:52.993124 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bb0216b-56a7-4822-bfce-0cf463a9d595" containerName="mariadb-database-create" Nov 24 08:15:52 crc kubenswrapper[4799]: I1124 08:15:52.993283 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="388f8341-6e2d-40e3-b35c-d1bd0d95e134" containerName="mariadb-account-create" Nov 24 08:15:52 crc kubenswrapper[4799]: I1124 08:15:52.994416 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7bnw6" Nov 24 08:15:52 crc kubenswrapper[4799]: I1124 08:15:52.999302 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.000173 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-xngkq" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.007161 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-7bnw6"] Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.044619 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrvj6\" (UniqueName: \"kubernetes.io/projected/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-kube-api-access-nrvj6\") pod \"heat-db-sync-7bnw6\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " pod="openstack/heat-db-sync-7bnw6" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.045038 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-combined-ca-bundle\") pod \"heat-db-sync-7bnw6\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " pod="openstack/heat-db-sync-7bnw6" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.045218 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-config-data\") pod \"heat-db-sync-7bnw6\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " pod="openstack/heat-db-sync-7bnw6" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.147346 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrvj6\" (UniqueName: \"kubernetes.io/projected/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-kube-api-access-nrvj6\") pod \"heat-db-sync-7bnw6\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " pod="openstack/heat-db-sync-7bnw6" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.147423 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-combined-ca-bundle\") pod \"heat-db-sync-7bnw6\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " pod="openstack/heat-db-sync-7bnw6" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.147454 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-config-data\") pod \"heat-db-sync-7bnw6\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " pod="openstack/heat-db-sync-7bnw6" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.158030 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-combined-ca-bundle\") pod \"heat-db-sync-7bnw6\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " pod="openstack/heat-db-sync-7bnw6" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.161294 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-config-data\") pod \"heat-db-sync-7bnw6\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " pod="openstack/heat-db-sync-7bnw6" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.174512 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrvj6\" (UniqueName: \"kubernetes.io/projected/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-kube-api-access-nrvj6\") pod \"heat-db-sync-7bnw6\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " pod="openstack/heat-db-sync-7bnw6" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.319338 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7bnw6" Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.874703 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-7bnw6"] Nov 24 08:15:53 crc kubenswrapper[4799]: W1124 08:15:53.887559 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6acabb45_e7c1_447f_aaaa_2a3514e3a80d.slice/crio-02d6b151400f66193bc3d09d6514e5293a1af52ba00aafe86bfee2ad5bd1be4a WatchSource:0}: Error finding container 02d6b151400f66193bc3d09d6514e5293a1af52ba00aafe86bfee2ad5bd1be4a: Status 404 returned error can't find the container with id 02d6b151400f66193bc3d09d6514e5293a1af52ba00aafe86bfee2ad5bd1be4a Nov 24 08:15:53 crc kubenswrapper[4799]: I1124 08:15:53.935688 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7bnw6" event={"ID":"6acabb45-e7c1-447f-aaaa-2a3514e3a80d","Type":"ContainerStarted","Data":"02d6b151400f66193bc3d09d6514e5293a1af52ba00aafe86bfee2ad5bd1be4a"} Nov 24 08:15:57 crc kubenswrapper[4799]: I1124 08:15:57.022105 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:57 crc kubenswrapper[4799]: I1124 08:15:57.023687 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:15:57 crc kubenswrapper[4799]: I1124 08:15:57.023803 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7cd9788b67-pld4l" podUID="75ed23b8-89f0-4527-bff2-5c8e60757c8f" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.98:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.98:8080: connect: connection refused" Nov 24 08:16:00 crc kubenswrapper[4799]: I1124 08:16:00.000377 4799 scope.go:117] "RemoveContainer" containerID="a33694f0c2326a842ebbf4f39ce982061a35e84b590d08b0db22f212c41a4fc2" Nov 24 08:16:03 crc kubenswrapper[4799]: I1124 08:16:03.055689 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7bnw6" event={"ID":"6acabb45-e7c1-447f-aaaa-2a3514e3a80d","Type":"ContainerStarted","Data":"b6510b4c8324aeec824855d1bef9711d745c4301afa6e642fa9998ee573fcc09"} Nov 24 08:16:03 crc kubenswrapper[4799]: I1124 08:16:03.072303 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-7bnw6" podStartSLOduration=2.333961443 podStartE2EDuration="11.072285397s" podCreationTimestamp="2025-11-24 08:15:52 +0000 UTC" firstStartedPulling="2025-11-24 08:15:53.89272665 +0000 UTC m=+5299.548709124" lastFinishedPulling="2025-11-24 08:16:02.631050614 +0000 UTC m=+5308.287033078" observedRunningTime="2025-11-24 08:16:03.069664322 +0000 UTC m=+5308.725646796" watchObservedRunningTime="2025-11-24 08:16:03.072285397 +0000 UTC m=+5308.728267871" Nov 24 08:16:05 crc kubenswrapper[4799]: I1124 08:16:05.079571 4799 generic.go:334] "Generic (PLEG): container finished" podID="6acabb45-e7c1-447f-aaaa-2a3514e3a80d" containerID="b6510b4c8324aeec824855d1bef9711d745c4301afa6e642fa9998ee573fcc09" exitCode=0 Nov 24 08:16:05 crc kubenswrapper[4799]: I1124 08:16:05.079688 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7bnw6" event={"ID":"6acabb45-e7c1-447f-aaaa-2a3514e3a80d","Type":"ContainerDied","Data":"b6510b4c8324aeec824855d1bef9711d745c4301afa6e642fa9998ee573fcc09"} Nov 24 08:16:06 crc kubenswrapper[4799]: I1124 08:16:06.459430 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7bnw6" Nov 24 08:16:06 crc kubenswrapper[4799]: I1124 08:16:06.658338 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-combined-ca-bundle\") pod \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " Nov 24 08:16:06 crc kubenswrapper[4799]: I1124 08:16:06.658554 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrvj6\" (UniqueName: \"kubernetes.io/projected/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-kube-api-access-nrvj6\") pod \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " Nov 24 08:16:06 crc kubenswrapper[4799]: I1124 08:16:06.658661 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-config-data\") pod \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\" (UID: \"6acabb45-e7c1-447f-aaaa-2a3514e3a80d\") " Nov 24 08:16:06 crc kubenswrapper[4799]: I1124 08:16:06.674052 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-kube-api-access-nrvj6" (OuterVolumeSpecName: "kube-api-access-nrvj6") pod "6acabb45-e7c1-447f-aaaa-2a3514e3a80d" (UID: "6acabb45-e7c1-447f-aaaa-2a3514e3a80d"). InnerVolumeSpecName "kube-api-access-nrvj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:16:06 crc kubenswrapper[4799]: I1124 08:16:06.690653 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6acabb45-e7c1-447f-aaaa-2a3514e3a80d" (UID: "6acabb45-e7c1-447f-aaaa-2a3514e3a80d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:16:06 crc kubenswrapper[4799]: I1124 08:16:06.744358 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-config-data" (OuterVolumeSpecName: "config-data") pod "6acabb45-e7c1-447f-aaaa-2a3514e3a80d" (UID: "6acabb45-e7c1-447f-aaaa-2a3514e3a80d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:16:06 crc kubenswrapper[4799]: I1124 08:16:06.761909 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:16:06 crc kubenswrapper[4799]: I1124 08:16:06.761967 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrvj6\" (UniqueName: \"kubernetes.io/projected/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-kube-api-access-nrvj6\") on node \"crc\" DevicePath \"\"" Nov 24 08:16:06 crc kubenswrapper[4799]: I1124 08:16:06.761981 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6acabb45-e7c1-447f-aaaa-2a3514e3a80d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:16:07 crc kubenswrapper[4799]: I1124 08:16:07.103405 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7bnw6" event={"ID":"6acabb45-e7c1-447f-aaaa-2a3514e3a80d","Type":"ContainerDied","Data":"02d6b151400f66193bc3d09d6514e5293a1af52ba00aafe86bfee2ad5bd1be4a"} Nov 24 08:16:07 crc kubenswrapper[4799]: I1124 08:16:07.103821 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02d6b151400f66193bc3d09d6514e5293a1af52ba00aafe86bfee2ad5bd1be4a" Nov 24 08:16:07 crc kubenswrapper[4799]: I1124 08:16:07.103539 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7bnw6" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.345443 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6ddc665688-m6glq"] Nov 24 08:16:08 crc kubenswrapper[4799]: E1124 08:16:08.346295 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6acabb45-e7c1-447f-aaaa-2a3514e3a80d" containerName="heat-db-sync" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.346315 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6acabb45-e7c1-447f-aaaa-2a3514e3a80d" containerName="heat-db-sync" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.346538 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6acabb45-e7c1-447f-aaaa-2a3514e3a80d" containerName="heat-db-sync" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.347364 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.352087 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.352177 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.357768 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-xngkq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.368095 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6ddc665688-m6glq"] Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.399611 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c6516-aca5-4f29-80da-3921fbb7728f-config-data\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.399693 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c6516-aca5-4f29-80da-3921fbb7728f-combined-ca-bundle\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.400138 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nllhw\" (UniqueName: \"kubernetes.io/projected/188c6516-aca5-4f29-80da-3921fbb7728f-kube-api-access-nllhw\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.400281 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/188c6516-aca5-4f29-80da-3921fbb7728f-config-data-custom\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.455261 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-65f685584f-5fhns"] Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.456483 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.464039 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.469120 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-57cccd9dc4-tq5ch"] Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.470768 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.473114 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.481073 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-57cccd9dc4-tq5ch"] Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.502430 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-config-data-custom\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.502508 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c6516-aca5-4f29-80da-3921fbb7728f-config-data\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.502608 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c6516-aca5-4f29-80da-3921fbb7728f-combined-ca-bundle\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.502743 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shl7w\" (UniqueName: \"kubernetes.io/projected/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-kube-api-access-shl7w\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.502803 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nllhw\" (UniqueName: \"kubernetes.io/projected/188c6516-aca5-4f29-80da-3921fbb7728f-kube-api-access-nllhw\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.502828 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-combined-ca-bundle\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.502874 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/188c6516-aca5-4f29-80da-3921fbb7728f-config-data-custom\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.502900 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-config-data\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.502933 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qxxw\" (UniqueName: \"kubernetes.io/projected/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-kube-api-access-5qxxw\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.503350 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-config-data-custom\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.503380 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-combined-ca-bundle\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.503445 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-config-data\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.512719 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c6516-aca5-4f29-80da-3921fbb7728f-config-data\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.513215 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c6516-aca5-4f29-80da-3921fbb7728f-combined-ca-bundle\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.518494 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/188c6516-aca5-4f29-80da-3921fbb7728f-config-data-custom\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.524954 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nllhw\" (UniqueName: \"kubernetes.io/projected/188c6516-aca5-4f29-80da-3921fbb7728f-kube-api-access-nllhw\") pod \"heat-engine-6ddc665688-m6glq\" (UID: \"188c6516-aca5-4f29-80da-3921fbb7728f\") " pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.551440 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-65f685584f-5fhns"] Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.605183 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-config-data\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.605280 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-config-data-custom\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.605366 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shl7w\" (UniqueName: \"kubernetes.io/projected/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-kube-api-access-shl7w\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.605437 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-combined-ca-bundle\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.605464 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-config-data\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.605495 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qxxw\" (UniqueName: \"kubernetes.io/projected/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-kube-api-access-5qxxw\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.605534 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-config-data-custom\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.605558 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-combined-ca-bundle\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.611835 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-config-data\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.614535 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-config-data-custom\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.614773 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-config-data\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.615441 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-combined-ca-bundle\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.621632 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-combined-ca-bundle\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.623593 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-config-data-custom\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.626381 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shl7w\" (UniqueName: \"kubernetes.io/projected/5fbab1f1-fb2a-4d63-9f7f-196b57104f8b-kube-api-access-shl7w\") pod \"heat-api-57cccd9dc4-tq5ch\" (UID: \"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b\") " pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.631518 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qxxw\" (UniqueName: \"kubernetes.io/projected/f02d4c9b-ef9f-4109-ae64-0fe7e07e321c-kube-api-access-5qxxw\") pod \"heat-cfnapi-65f685584f-5fhns\" (UID: \"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c\") " pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.681451 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.787419 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:08 crc kubenswrapper[4799]: I1124 08:16:08.811797 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:09 crc kubenswrapper[4799]: I1124 08:16:09.282976 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6ddc665688-m6glq"] Nov 24 08:16:09 crc kubenswrapper[4799]: I1124 08:16:09.381822 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-65f685584f-5fhns"] Nov 24 08:16:09 crc kubenswrapper[4799]: I1124 08:16:09.532906 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:16:09 crc kubenswrapper[4799]: I1124 08:16:09.647332 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-57cccd9dc4-tq5ch"] Nov 24 08:16:09 crc kubenswrapper[4799]: W1124 08:16:09.665512 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fbab1f1_fb2a_4d63_9f7f_196b57104f8b.slice/crio-56a8045afe6a4fb0156eb533df8257cdd52fd9538cb5ce183f07fb891e57e698 WatchSource:0}: Error finding container 56a8045afe6a4fb0156eb533df8257cdd52fd9538cb5ce183f07fb891e57e698: Status 404 returned error can't find the container with id 56a8045afe6a4fb0156eb533df8257cdd52fd9538cb5ce183f07fb891e57e698 Nov 24 08:16:10 crc kubenswrapper[4799]: I1124 08:16:10.144300 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-65f685584f-5fhns" event={"ID":"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c","Type":"ContainerStarted","Data":"e0c514ff3fe5f40fa99455f82234aa36d1cfafb41ab4d27bb27023ddf8f54fcf"} Nov 24 08:16:10 crc kubenswrapper[4799]: I1124 08:16:10.146812 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-57cccd9dc4-tq5ch" event={"ID":"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b","Type":"ContainerStarted","Data":"56a8045afe6a4fb0156eb533df8257cdd52fd9538cb5ce183f07fb891e57e698"} Nov 24 08:16:10 crc kubenswrapper[4799]: I1124 08:16:10.149568 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6ddc665688-m6glq" event={"ID":"188c6516-aca5-4f29-80da-3921fbb7728f","Type":"ContainerStarted","Data":"1ab700138b5accad823b70e07ae0bdb32cdb533b31a054865a35d0dffa1df025"} Nov 24 08:16:10 crc kubenswrapper[4799]: I1124 08:16:10.149644 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6ddc665688-m6glq" event={"ID":"188c6516-aca5-4f29-80da-3921fbb7728f","Type":"ContainerStarted","Data":"0a6d06ba478a81e58915b91d6b620783f2165e00ecbd5b0121ea337c465d0a50"} Nov 24 08:16:10 crc kubenswrapper[4799]: I1124 08:16:10.149763 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:10 crc kubenswrapper[4799]: I1124 08:16:10.171050 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6ddc665688-m6glq" podStartSLOduration=2.171029479 podStartE2EDuration="2.171029479s" podCreationTimestamp="2025-11-24 08:16:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:10.165209264 +0000 UTC m=+5315.821191738" watchObservedRunningTime="2025-11-24 08:16:10.171029479 +0000 UTC m=+5315.827012263" Nov 24 08:16:11 crc kubenswrapper[4799]: I1124 08:16:11.043572 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-d14e-account-create-9lhsh"] Nov 24 08:16:11 crc kubenswrapper[4799]: I1124 08:16:11.058394 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-dtcc5"] Nov 24 08:16:11 crc kubenswrapper[4799]: I1124 08:16:11.069913 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-d14e-account-create-9lhsh"] Nov 24 08:16:11 crc kubenswrapper[4799]: I1124 08:16:11.076746 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-dtcc5"] Nov 24 08:16:11 crc kubenswrapper[4799]: I1124 08:16:11.639793 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c5723e9-24ac-4cfa-b181-7030c18c2d27" path="/var/lib/kubelet/pods/4c5723e9-24ac-4cfa-b181-7030c18c2d27/volumes" Nov 24 08:16:11 crc kubenswrapper[4799]: I1124 08:16:11.641158 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="508bd06c-03d1-44fe-9425-5a36f40a60b5" path="/var/lib/kubelet/pods/508bd06c-03d1-44fe-9425-5a36f40a60b5/volumes" Nov 24 08:16:11 crc kubenswrapper[4799]: I1124 08:16:11.646503 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7cd9788b67-pld4l" Nov 24 08:16:11 crc kubenswrapper[4799]: I1124 08:16:11.753607 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6896b5c567-jgd8d"] Nov 24 08:16:11 crc kubenswrapper[4799]: I1124 08:16:11.753933 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6896b5c567-jgd8d" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon-log" containerID="cri-o://71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744" gracePeriod=30 Nov 24 08:16:11 crc kubenswrapper[4799]: I1124 08:16:11.754466 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6896b5c567-jgd8d" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon" containerID="cri-o://82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c" gracePeriod=30 Nov 24 08:16:12 crc kubenswrapper[4799]: I1124 08:16:12.184970 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-65f685584f-5fhns" event={"ID":"f02d4c9b-ef9f-4109-ae64-0fe7e07e321c","Type":"ContainerStarted","Data":"7f153b67cbbdeeb87a928b0af07045bd5e59e4757453414197e7d512d6fa06ad"} Nov 24 08:16:12 crc kubenswrapper[4799]: I1124 08:16:12.185026 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:12 crc kubenswrapper[4799]: I1124 08:16:12.186529 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-57cccd9dc4-tq5ch" event={"ID":"5fbab1f1-fb2a-4d63-9f7f-196b57104f8b","Type":"ContainerStarted","Data":"01c5bc1d6a671b7f3074412690461a87b1df3671b03d9884bb797d2f7d489f97"} Nov 24 08:16:12 crc kubenswrapper[4799]: I1124 08:16:12.186676 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:12 crc kubenswrapper[4799]: I1124 08:16:12.201759 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-65f685584f-5fhns" podStartSLOduration=2.373718055 podStartE2EDuration="4.201744019s" podCreationTimestamp="2025-11-24 08:16:08 +0000 UTC" firstStartedPulling="2025-11-24 08:16:09.384951481 +0000 UTC m=+5315.040933945" lastFinishedPulling="2025-11-24 08:16:11.212977435 +0000 UTC m=+5316.868959909" observedRunningTime="2025-11-24 08:16:12.199116725 +0000 UTC m=+5317.855099199" watchObservedRunningTime="2025-11-24 08:16:12.201744019 +0000 UTC m=+5317.857726493" Nov 24 08:16:15 crc kubenswrapper[4799]: I1124 08:16:15.222706 4799 generic.go:334] "Generic (PLEG): container finished" podID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerID="82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c" exitCode=0 Nov 24 08:16:15 crc kubenswrapper[4799]: I1124 08:16:15.222810 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6896b5c567-jgd8d" event={"ID":"700b8e45-d105-4ca5-b89c-0c7c0171c0fb","Type":"ContainerDied","Data":"82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c"} Nov 24 08:16:20 crc kubenswrapper[4799]: I1124 08:16:20.104720 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6896b5c567-jgd8d" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.92:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.92:8080: connect: connection refused" Nov 24 08:16:20 crc kubenswrapper[4799]: I1124 08:16:20.105257 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-65f685584f-5fhns" Nov 24 08:16:20 crc kubenswrapper[4799]: I1124 08:16:20.127776 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-57cccd9dc4-tq5ch" podStartSLOduration=10.584509786 podStartE2EDuration="12.12775121s" podCreationTimestamp="2025-11-24 08:16:08 +0000 UTC" firstStartedPulling="2025-11-24 08:16:09.67287323 +0000 UTC m=+5315.328855704" lastFinishedPulling="2025-11-24 08:16:11.216114654 +0000 UTC m=+5316.872097128" observedRunningTime="2025-11-24 08:16:12.219783202 +0000 UTC m=+5317.875765676" watchObservedRunningTime="2025-11-24 08:16:20.12775121 +0000 UTC m=+5325.783733704" Nov 24 08:16:20 crc kubenswrapper[4799]: I1124 08:16:20.206328 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-57cccd9dc4-tq5ch" Nov 24 08:16:21 crc kubenswrapper[4799]: I1124 08:16:21.028821 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-sw84r"] Nov 24 08:16:21 crc kubenswrapper[4799]: I1124 08:16:21.037208 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-sw84r"] Nov 24 08:16:21 crc kubenswrapper[4799]: I1124 08:16:21.643416 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4" path="/var/lib/kubelet/pods/d8f0d2ed-dc7a-4b05-bd5b-e6ee03af1ba4/volumes" Nov 24 08:16:28 crc kubenswrapper[4799]: I1124 08:16:28.712192 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6ddc665688-m6glq" Nov 24 08:16:30 crc kubenswrapper[4799]: I1124 08:16:30.103834 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6896b5c567-jgd8d" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.92:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.92:8080: connect: connection refused" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.140189 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x"] Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.143872 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.145728 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.165425 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x"] Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.333550 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.334004 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.334082 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvdpc\" (UniqueName: \"kubernetes.io/projected/b051e98e-b756-4df6-ba65-29e4ac6572ca-kube-api-access-wvdpc\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.435552 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.435656 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.435684 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvdpc\" (UniqueName: \"kubernetes.io/projected/b051e98e-b756-4df6-ba65-29e4ac6572ca-kube-api-access-wvdpc\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.436341 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.436372 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.463282 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvdpc\" (UniqueName: \"kubernetes.io/projected/b051e98e-b756-4df6-ba65-29e4ac6572ca-kube-api-access-wvdpc\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.484598 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:38 crc kubenswrapper[4799]: I1124 08:16:38.945676 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x"] Nov 24 08:16:39 crc kubenswrapper[4799]: I1124 08:16:39.482395 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" event={"ID":"b051e98e-b756-4df6-ba65-29e4ac6572ca","Type":"ContainerStarted","Data":"de834268cb98c25c95affffc761ae016fb00d60be3c1c2895a92d15b73eb8d4e"} Nov 24 08:16:39 crc kubenswrapper[4799]: I1124 08:16:39.484900 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" event={"ID":"b051e98e-b756-4df6-ba65-29e4ac6572ca","Type":"ContainerStarted","Data":"916142c15d95a6433c576515e674e550684bbc95f8596c4433ac731fb4a6d034"} Nov 24 08:16:40 crc kubenswrapper[4799]: I1124 08:16:40.105682 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6896b5c567-jgd8d" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.92:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.92:8080: connect: connection refused" Nov 24 08:16:40 crc kubenswrapper[4799]: I1124 08:16:40.105803 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:16:40 crc kubenswrapper[4799]: I1124 08:16:40.506488 4799 generic.go:334] "Generic (PLEG): container finished" podID="b051e98e-b756-4df6-ba65-29e4ac6572ca" containerID="de834268cb98c25c95affffc761ae016fb00d60be3c1c2895a92d15b73eb8d4e" exitCode=0 Nov 24 08:16:40 crc kubenswrapper[4799]: I1124 08:16:40.506566 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" event={"ID":"b051e98e-b756-4df6-ba65-29e4ac6572ca","Type":"ContainerDied","Data":"de834268cb98c25c95affffc761ae016fb00d60be3c1c2895a92d15b73eb8d4e"} Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.299292 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.428033 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-logs\") pod \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.428103 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9wj6\" (UniqueName: \"kubernetes.io/projected/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-kube-api-access-w9wj6\") pod \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.428176 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-horizon-secret-key\") pod \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.428243 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-config-data\") pod \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.428289 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-scripts\") pod \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\" (UID: \"700b8e45-d105-4ca5-b89c-0c7c0171c0fb\") " Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.430209 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-logs" (OuterVolumeSpecName: "logs") pod "700b8e45-d105-4ca5-b89c-0c7c0171c0fb" (UID: "700b8e45-d105-4ca5-b89c-0c7c0171c0fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.443889 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-kube-api-access-w9wj6" (OuterVolumeSpecName: "kube-api-access-w9wj6") pod "700b8e45-d105-4ca5-b89c-0c7c0171c0fb" (UID: "700b8e45-d105-4ca5-b89c-0c7c0171c0fb"). InnerVolumeSpecName "kube-api-access-w9wj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.457175 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "700b8e45-d105-4ca5-b89c-0c7c0171c0fb" (UID: "700b8e45-d105-4ca5-b89c-0c7c0171c0fb"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.493457 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-scripts" (OuterVolumeSpecName: "scripts") pod "700b8e45-d105-4ca5-b89c-0c7c0171c0fb" (UID: "700b8e45-d105-4ca5-b89c-0c7c0171c0fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.533906 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.533972 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9wj6\" (UniqueName: \"kubernetes.io/projected/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-kube-api-access-w9wj6\") on node \"crc\" DevicePath \"\"" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.533989 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.533999 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.540419 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-config-data" (OuterVolumeSpecName: "config-data") pod "700b8e45-d105-4ca5-b89c-0c7c0171c0fb" (UID: "700b8e45-d105-4ca5-b89c-0c7c0171c0fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.566874 4799 generic.go:334] "Generic (PLEG): container finished" podID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerID="71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744" exitCode=137 Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.567172 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6896b5c567-jgd8d" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.567078 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6896b5c567-jgd8d" event={"ID":"700b8e45-d105-4ca5-b89c-0c7c0171c0fb","Type":"ContainerDied","Data":"71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744"} Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.567386 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6896b5c567-jgd8d" event={"ID":"700b8e45-d105-4ca5-b89c-0c7c0171c0fb","Type":"ContainerDied","Data":"9462e48835f5c7546f11e23bc846fe7531de5b26b9697a42aaaa91429451e0f4"} Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.567458 4799 scope.go:117] "RemoveContainer" containerID="82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.575546 4799 generic.go:334] "Generic (PLEG): container finished" podID="b051e98e-b756-4df6-ba65-29e4ac6572ca" containerID="5b378f5018210c5aad4ec79b30559995cccbe8a6639186fa485b3af49fd2cb12" exitCode=0 Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.575589 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" event={"ID":"b051e98e-b756-4df6-ba65-29e4ac6572ca","Type":"ContainerDied","Data":"5b378f5018210c5aad4ec79b30559995cccbe8a6639186fa485b3af49fd2cb12"} Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.633365 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6896b5c567-jgd8d"] Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.636064 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/700b8e45-d105-4ca5-b89c-0c7c0171c0fb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.642629 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6896b5c567-jgd8d"] Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.782812 4799 scope.go:117] "RemoveContainer" containerID="71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.829188 4799 scope.go:117] "RemoveContainer" containerID="82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c" Nov 24 08:16:42 crc kubenswrapper[4799]: E1124 08:16:42.830408 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c\": container with ID starting with 82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c not found: ID does not exist" containerID="82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.830459 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c"} err="failed to get container status \"82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c\": rpc error: code = NotFound desc = could not find container \"82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c\": container with ID starting with 82c38f82fade692b44fcd349d0447630507251aec5ad8f8fc348b4bbdd0d7c3c not found: ID does not exist" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.830487 4799 scope.go:117] "RemoveContainer" containerID="71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744" Nov 24 08:16:42 crc kubenswrapper[4799]: E1124 08:16:42.831196 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744\": container with ID starting with 71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744 not found: ID does not exist" containerID="71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744" Nov 24 08:16:42 crc kubenswrapper[4799]: I1124 08:16:42.831229 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744"} err="failed to get container status \"71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744\": rpc error: code = NotFound desc = could not find container \"71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744\": container with ID starting with 71979b15f6c6ec3210817cd1e61b37af2881a68ae019fbeffab0d466680d9744 not found: ID does not exist" Nov 24 08:16:43 crc kubenswrapper[4799]: I1124 08:16:43.592102 4799 generic.go:334] "Generic (PLEG): container finished" podID="b051e98e-b756-4df6-ba65-29e4ac6572ca" containerID="8a5ff33c3ab61b216092e1d46008e02ea30791eee78d5fb29e71117818baea6c" exitCode=0 Nov 24 08:16:43 crc kubenswrapper[4799]: I1124 08:16:43.592159 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" event={"ID":"b051e98e-b756-4df6-ba65-29e4ac6572ca","Type":"ContainerDied","Data":"8a5ff33c3ab61b216092e1d46008e02ea30791eee78d5fb29e71117818baea6c"} Nov 24 08:16:43 crc kubenswrapper[4799]: I1124 08:16:43.641904 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" path="/var/lib/kubelet/pods/700b8e45-d105-4ca5-b89c-0c7c0171c0fb/volumes" Nov 24 08:16:44 crc kubenswrapper[4799]: I1124 08:16:44.972593 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.090518 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-util\") pod \"b051e98e-b756-4df6-ba65-29e4ac6572ca\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.090598 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvdpc\" (UniqueName: \"kubernetes.io/projected/b051e98e-b756-4df6-ba65-29e4ac6572ca-kube-api-access-wvdpc\") pod \"b051e98e-b756-4df6-ba65-29e4ac6572ca\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.090762 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-bundle\") pod \"b051e98e-b756-4df6-ba65-29e4ac6572ca\" (UID: \"b051e98e-b756-4df6-ba65-29e4ac6572ca\") " Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.092934 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-bundle" (OuterVolumeSpecName: "bundle") pod "b051e98e-b756-4df6-ba65-29e4ac6572ca" (UID: "b051e98e-b756-4df6-ba65-29e4ac6572ca"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.096626 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b051e98e-b756-4df6-ba65-29e4ac6572ca-kube-api-access-wvdpc" (OuterVolumeSpecName: "kube-api-access-wvdpc") pod "b051e98e-b756-4df6-ba65-29e4ac6572ca" (UID: "b051e98e-b756-4df6-ba65-29e4ac6572ca"). InnerVolumeSpecName "kube-api-access-wvdpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.108444 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-util" (OuterVolumeSpecName: "util") pod "b051e98e-b756-4df6-ba65-29e4ac6572ca" (UID: "b051e98e-b756-4df6-ba65-29e4ac6572ca"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.192888 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvdpc\" (UniqueName: \"kubernetes.io/projected/b051e98e-b756-4df6-ba65-29e4ac6572ca-kube-api-access-wvdpc\") on node \"crc\" DevicePath \"\"" Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.192948 4799 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.192968 4799 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b051e98e-b756-4df6-ba65-29e4ac6572ca-util\") on node \"crc\" DevicePath \"\"" Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.619119 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" event={"ID":"b051e98e-b756-4df6-ba65-29e4ac6572ca","Type":"ContainerDied","Data":"916142c15d95a6433c576515e674e550684bbc95f8596c4433ac731fb4a6d034"} Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.619192 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="916142c15d95a6433c576515e674e550684bbc95f8596c4433ac731fb4a6d034" Nov 24 08:16:45 crc kubenswrapper[4799]: I1124 08:16:45.619207 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x" Nov 24 08:16:50 crc kubenswrapper[4799]: I1124 08:16:50.081761 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-3e6b-account-create-48czt"] Nov 24 08:16:50 crc kubenswrapper[4799]: I1124 08:16:50.090917 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-lpbqc"] Nov 24 08:16:50 crc kubenswrapper[4799]: I1124 08:16:50.100258 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-3e6b-account-create-48czt"] Nov 24 08:16:50 crc kubenswrapper[4799]: I1124 08:16:50.108413 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-lpbqc"] Nov 24 08:16:51 crc kubenswrapper[4799]: I1124 08:16:51.637981 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3de7dad8-7eca-4d02-ac2f-8d07a3608d75" path="/var/lib/kubelet/pods/3de7dad8-7eca-4d02-ac2f-8d07a3608d75/volumes" Nov 24 08:16:51 crc kubenswrapper[4799]: I1124 08:16:51.638537 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96d54256-8e3b-4ccb-96b5-39bad6ebf371" path="/var/lib/kubelet/pods/96d54256-8e3b-4ccb-96b5-39bad6ebf371/volumes" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.318251 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-ftn2m"] Nov 24 08:16:56 crc kubenswrapper[4799]: E1124 08:16:56.320026 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b051e98e-b756-4df6-ba65-29e4ac6572ca" containerName="util" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.320112 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b051e98e-b756-4df6-ba65-29e4ac6572ca" containerName="util" Nov 24 08:16:56 crc kubenswrapper[4799]: E1124 08:16:56.320198 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b051e98e-b756-4df6-ba65-29e4ac6572ca" containerName="pull" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.320270 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b051e98e-b756-4df6-ba65-29e4ac6572ca" containerName="pull" Nov 24 08:16:56 crc kubenswrapper[4799]: E1124 08:16:56.320329 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b051e98e-b756-4df6-ba65-29e4ac6572ca" containerName="extract" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.320387 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b051e98e-b756-4df6-ba65-29e4ac6572ca" containerName="extract" Nov 24 08:16:56 crc kubenswrapper[4799]: E1124 08:16:56.320462 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon-log" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.320518 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon-log" Nov 24 08:16:56 crc kubenswrapper[4799]: E1124 08:16:56.320589 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.320645 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.320918 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b051e98e-b756-4df6-ba65-29e4ac6572ca" containerName="extract" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.320991 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon-log" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.321062 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="700b8e45-d105-4ca5-b89c-0c7c0171c0fb" containerName="horizon" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.322723 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ftn2m" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.324699 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.326000 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-xmp7x" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.326291 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.330673 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-ftn2m"] Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.440211 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2"] Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.441778 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.444516 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.447577 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-b9k4q" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.452010 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl"] Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.453358 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.459310 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgjws\" (UniqueName: \"kubernetes.io/projected/f6f9c83e-09c2-43c5-bb06-a0ea47981131-kube-api-access-jgjws\") pod \"obo-prometheus-operator-668cf9dfbb-ftn2m\" (UID: \"f6f9c83e-09c2-43c5-bb06-a0ea47981131\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ftn2m" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.463024 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2"] Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.485695 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl"] Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.564440 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5ac7f1a5-756a-4dab-bde7-a68e99f05ed1-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-vb9r2\" (UID: \"5ac7f1a5-756a-4dab-bde7-a68e99f05ed1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.564540 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5ac7f1a5-756a-4dab-bde7-a68e99f05ed1-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-vb9r2\" (UID: \"5ac7f1a5-756a-4dab-bde7-a68e99f05ed1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.564581 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b438aea-0c2b-41eb-8e2d-8b77fa4a7574-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-j9twl\" (UID: \"0b438aea-0c2b-41eb-8e2d-8b77fa4a7574\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.564689 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b438aea-0c2b-41eb-8e2d-8b77fa4a7574-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-j9twl\" (UID: \"0b438aea-0c2b-41eb-8e2d-8b77fa4a7574\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.564802 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgjws\" (UniqueName: \"kubernetes.io/projected/f6f9c83e-09c2-43c5-bb06-a0ea47981131-kube-api-access-jgjws\") pod \"obo-prometheus-operator-668cf9dfbb-ftn2m\" (UID: \"f6f9c83e-09c2-43c5-bb06-a0ea47981131\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ftn2m" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.585573 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgjws\" (UniqueName: \"kubernetes.io/projected/f6f9c83e-09c2-43c5-bb06-a0ea47981131-kube-api-access-jgjws\") pod \"obo-prometheus-operator-668cf9dfbb-ftn2m\" (UID: \"f6f9c83e-09c2-43c5-bb06-a0ea47981131\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ftn2m" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.645039 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ftn2m" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.649566 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-wnwkw"] Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.651287 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.655732 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.659050 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-cqmtx" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.667843 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b438aea-0c2b-41eb-8e2d-8b77fa4a7574-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-j9twl\" (UID: \"0b438aea-0c2b-41eb-8e2d-8b77fa4a7574\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.668188 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4wt8\" (UniqueName: \"kubernetes.io/projected/3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820-kube-api-access-g4wt8\") pod \"observability-operator-d8bb48f5d-wnwkw\" (UID: \"3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820\") " pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.668586 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b438aea-0c2b-41eb-8e2d-8b77fa4a7574-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-j9twl\" (UID: \"0b438aea-0c2b-41eb-8e2d-8b77fa4a7574\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.668717 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-wnwkw\" (UID: \"3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820\") " pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.668944 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5ac7f1a5-756a-4dab-bde7-a68e99f05ed1-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-vb9r2\" (UID: \"5ac7f1a5-756a-4dab-bde7-a68e99f05ed1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.669141 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5ac7f1a5-756a-4dab-bde7-a68e99f05ed1-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-vb9r2\" (UID: \"5ac7f1a5-756a-4dab-bde7-a68e99f05ed1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.668727 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-wnwkw"] Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.679692 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b438aea-0c2b-41eb-8e2d-8b77fa4a7574-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-j9twl\" (UID: \"0b438aea-0c2b-41eb-8e2d-8b77fa4a7574\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.681454 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5ac7f1a5-756a-4dab-bde7-a68e99f05ed1-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-vb9r2\" (UID: \"5ac7f1a5-756a-4dab-bde7-a68e99f05ed1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.698481 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b438aea-0c2b-41eb-8e2d-8b77fa4a7574-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-j9twl\" (UID: \"0b438aea-0c2b-41eb-8e2d-8b77fa4a7574\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.707291 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5ac7f1a5-756a-4dab-bde7-a68e99f05ed1-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86db4855d-vb9r2\" (UID: \"5ac7f1a5-756a-4dab-bde7-a68e99f05ed1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.753821 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-26xdx"] Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.758267 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.767457 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-26xdx" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.770711 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4wt8\" (UniqueName: \"kubernetes.io/projected/3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820-kube-api-access-g4wt8\") pod \"observability-operator-d8bb48f5d-wnwkw\" (UID: \"3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820\") " pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.770768 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4tpv\" (UniqueName: \"kubernetes.io/projected/26315ae1-10ac-4122-b2ac-6ae36c41fd84-kube-api-access-x4tpv\") pod \"perses-operator-5446b9c989-26xdx\" (UID: \"26315ae1-10ac-4122-b2ac-6ae36c41fd84\") " pod="openshift-operators/perses-operator-5446b9c989-26xdx" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.770802 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/26315ae1-10ac-4122-b2ac-6ae36c41fd84-openshift-service-ca\") pod \"perses-operator-5446b9c989-26xdx\" (UID: \"26315ae1-10ac-4122-b2ac-6ae36c41fd84\") " pod="openshift-operators/perses-operator-5446b9c989-26xdx" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.770880 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-wnwkw\" (UID: \"3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820\") " pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.778320 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.779877 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-wnwkw\" (UID: \"3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820\") " pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.780255 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-b7md6" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.815172 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4wt8\" (UniqueName: \"kubernetes.io/projected/3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820-kube-api-access-g4wt8\") pod \"observability-operator-d8bb48f5d-wnwkw\" (UID: \"3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820\") " pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.822860 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-26xdx"] Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.866496 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.873478 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4tpv\" (UniqueName: \"kubernetes.io/projected/26315ae1-10ac-4122-b2ac-6ae36c41fd84-kube-api-access-x4tpv\") pod \"perses-operator-5446b9c989-26xdx\" (UID: \"26315ae1-10ac-4122-b2ac-6ae36c41fd84\") " pod="openshift-operators/perses-operator-5446b9c989-26xdx" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.873549 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/26315ae1-10ac-4122-b2ac-6ae36c41fd84-openshift-service-ca\") pod \"perses-operator-5446b9c989-26xdx\" (UID: \"26315ae1-10ac-4122-b2ac-6ae36c41fd84\") " pod="openshift-operators/perses-operator-5446b9c989-26xdx" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.875205 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/26315ae1-10ac-4122-b2ac-6ae36c41fd84-openshift-service-ca\") pod \"perses-operator-5446b9c989-26xdx\" (UID: \"26315ae1-10ac-4122-b2ac-6ae36c41fd84\") " pod="openshift-operators/perses-operator-5446b9c989-26xdx" Nov 24 08:16:56 crc kubenswrapper[4799]: I1124 08:16:56.905662 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4tpv\" (UniqueName: \"kubernetes.io/projected/26315ae1-10ac-4122-b2ac-6ae36c41fd84-kube-api-access-x4tpv\") pod \"perses-operator-5446b9c989-26xdx\" (UID: \"26315ae1-10ac-4122-b2ac-6ae36c41fd84\") " pod="openshift-operators/perses-operator-5446b9c989-26xdx" Nov 24 08:16:57 crc kubenswrapper[4799]: I1124 08:16:57.181555 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-26xdx" Nov 24 08:16:57 crc kubenswrapper[4799]: I1124 08:16:57.324656 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-ftn2m"] Nov 24 08:16:57 crc kubenswrapper[4799]: I1124 08:16:57.486573 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2"] Nov 24 08:16:57 crc kubenswrapper[4799]: I1124 08:16:57.508759 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl"] Nov 24 08:16:57 crc kubenswrapper[4799]: W1124 08:16:57.514070 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b438aea_0c2b_41eb_8e2d_8b77fa4a7574.slice/crio-ba138e9f9802a4337a22739006418001d736b454b5611d8e62f64fcd37eb3f90 WatchSource:0}: Error finding container ba138e9f9802a4337a22739006418001d736b454b5611d8e62f64fcd37eb3f90: Status 404 returned error can't find the container with id ba138e9f9802a4337a22739006418001d736b454b5611d8e62f64fcd37eb3f90 Nov 24 08:16:57 crc kubenswrapper[4799]: I1124 08:16:57.704478 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-wnwkw"] Nov 24 08:16:57 crc kubenswrapper[4799]: I1124 08:16:57.791820 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ftn2m" event={"ID":"f6f9c83e-09c2-43c5-bb06-a0ea47981131","Type":"ContainerStarted","Data":"8c82b6142bce005a83fc1d71d6ad552ba1be097aef70815730e5ac20531ae98d"} Nov 24 08:16:57 crc kubenswrapper[4799]: I1124 08:16:57.797410 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" event={"ID":"3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820","Type":"ContainerStarted","Data":"e1bc195b6adef867d152a8466cab383b77d14a0f5debd236830485eede567e9d"} Nov 24 08:16:57 crc kubenswrapper[4799]: I1124 08:16:57.814143 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2" event={"ID":"5ac7f1a5-756a-4dab-bde7-a68e99f05ed1","Type":"ContainerStarted","Data":"34f3a9e47371ae00448ec820436de07bc9098686b50a85e9d4100613b0f92d29"} Nov 24 08:16:57 crc kubenswrapper[4799]: I1124 08:16:57.821762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl" event={"ID":"0b438aea-0c2b-41eb-8e2d-8b77fa4a7574","Type":"ContainerStarted","Data":"ba138e9f9802a4337a22739006418001d736b454b5611d8e62f64fcd37eb3f90"} Nov 24 08:16:57 crc kubenswrapper[4799]: I1124 08:16:57.831355 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-26xdx"] Nov 24 08:16:58 crc kubenswrapper[4799]: I1124 08:16:58.844600 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-26xdx" event={"ID":"26315ae1-10ac-4122-b2ac-6ae36c41fd84","Type":"ContainerStarted","Data":"aad5e93db20ecd7fe16a803e8e37be601379205ae6e66d306a5e0633327a1ef4"} Nov 24 08:16:59 crc kubenswrapper[4799]: I1124 08:16:59.117788 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-kcww7"] Nov 24 08:16:59 crc kubenswrapper[4799]: I1124 08:16:59.149232 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-kcww7"] Nov 24 08:16:59 crc kubenswrapper[4799]: I1124 08:16:59.644937 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8969b312-9957-41a5-a640-635f79eb16e4" path="/var/lib/kubelet/pods/8969b312-9957-41a5-a640-635f79eb16e4/volumes" Nov 24 08:17:02 crc kubenswrapper[4799]: I1124 08:17:02.633620 4799 scope.go:117] "RemoveContainer" containerID="0c2ec8aeaf7aa6a29cd3ef59e6b1315cbc94672034fa02675d5e4275282a8156" Nov 24 08:17:07 crc kubenswrapper[4799]: I1124 08:17:07.463402 4799 scope.go:117] "RemoveContainer" containerID="d572fa47053ae2af441fb1776d1b38095adb207b34655025aeeae3ca701fb086" Nov 24 08:17:07 crc kubenswrapper[4799]: I1124 08:17:07.537580 4799 scope.go:117] "RemoveContainer" containerID="01ea2798e6be3b95b69f465afe110ac272165a163aeeed1ee26cfa6510047bd6" Nov 24 08:17:07 crc kubenswrapper[4799]: I1124 08:17:07.965935 4799 scope.go:117] "RemoveContainer" containerID="38a0f8c178a84731da58580d7dfa58bef30c8740c5c85746ae760163e298b8d4" Nov 24 08:17:08 crc kubenswrapper[4799]: I1124 08:17:08.019637 4799 scope.go:117] "RemoveContainer" containerID="cb9401eac7115b5bcb4c506c146c4ee97d369100f86db3480ec84ff39aa0646a" Nov 24 08:17:08 crc kubenswrapper[4799]: I1124 08:17:08.074372 4799 scope.go:117] "RemoveContainer" containerID="ddb8609e7c37cf623fef37f8ed548cece157e70125cfd22887adb730d639cd6b" Nov 24 08:17:08 crc kubenswrapper[4799]: I1124 08:17:08.975562 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-26xdx" event={"ID":"26315ae1-10ac-4122-b2ac-6ae36c41fd84","Type":"ContainerStarted","Data":"eab8edb2f795e7341977093596640e78ff55889b0d1e6011a36e4d8898718fc2"} Nov 24 08:17:08 crc kubenswrapper[4799]: I1124 08:17:08.975875 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-26xdx" Nov 24 08:17:08 crc kubenswrapper[4799]: I1124 08:17:08.977623 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ftn2m" event={"ID":"f6f9c83e-09c2-43c5-bb06-a0ea47981131","Type":"ContainerStarted","Data":"bb13a14d0530974a1022758eb5fe2a062e3a08b156551353759fda28d934040c"} Nov 24 08:17:08 crc kubenswrapper[4799]: I1124 08:17:08.979334 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2" event={"ID":"5ac7f1a5-756a-4dab-bde7-a68e99f05ed1","Type":"ContainerStarted","Data":"b3bfb18b1f4e9316633a8b7c347d1ec07252dff2c51dadb9e1e02554eb8b9fb4"} Nov 24 08:17:08 crc kubenswrapper[4799]: I1124 08:17:08.981385 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl" event={"ID":"0b438aea-0c2b-41eb-8e2d-8b77fa4a7574","Type":"ContainerStarted","Data":"cb78d1c80b459f00950fe89f522725acc7c0e4ba511d1c2302cd8bd715077f91"} Nov 24 08:17:08 crc kubenswrapper[4799]: I1124 08:17:08.982957 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" event={"ID":"3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820","Type":"ContainerStarted","Data":"06d45d0ce0d3edf93ebcf40ecf2e6d79297179e48e8cacee304b54e269314be8"} Nov 24 08:17:08 crc kubenswrapper[4799]: I1124 08:17:08.983433 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" Nov 24 08:17:08 crc kubenswrapper[4799]: I1124 08:17:08.987174 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" Nov 24 08:17:09 crc kubenswrapper[4799]: I1124 08:17:09.001255 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-26xdx" podStartSLOduration=3.372005738 podStartE2EDuration="13.001233907s" podCreationTimestamp="2025-11-24 08:16:56 +0000 UTC" firstStartedPulling="2025-11-24 08:16:57.839706334 +0000 UTC m=+5363.495688798" lastFinishedPulling="2025-11-24 08:17:07.468934493 +0000 UTC m=+5373.124916967" observedRunningTime="2025-11-24 08:17:08.994046803 +0000 UTC m=+5374.650029277" watchObservedRunningTime="2025-11-24 08:17:09.001233907 +0000 UTC m=+5374.657216381" Nov 24 08:17:09 crc kubenswrapper[4799]: I1124 08:17:09.026174 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-ftn2m" podStartSLOduration=2.898891439 podStartE2EDuration="13.026155565s" podCreationTimestamp="2025-11-24 08:16:56 +0000 UTC" firstStartedPulling="2025-11-24 08:16:57.342045538 +0000 UTC m=+5362.998028012" lastFinishedPulling="2025-11-24 08:17:07.469309664 +0000 UTC m=+5373.125292138" observedRunningTime="2025-11-24 08:17:09.016033307 +0000 UTC m=+5374.672015791" watchObservedRunningTime="2025-11-24 08:17:09.026155565 +0000 UTC m=+5374.682138039" Nov 24 08:17:09 crc kubenswrapper[4799]: I1124 08:17:09.044709 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-vb9r2" podStartSLOduration=3.081587449 podStartE2EDuration="13.044691161s" podCreationTimestamp="2025-11-24 08:16:56 +0000 UTC" firstStartedPulling="2025-11-24 08:16:57.500456608 +0000 UTC m=+5363.156439082" lastFinishedPulling="2025-11-24 08:17:07.46356032 +0000 UTC m=+5373.119542794" observedRunningTime="2025-11-24 08:17:09.043574429 +0000 UTC m=+5374.699556903" watchObservedRunningTime="2025-11-24 08:17:09.044691161 +0000 UTC m=+5374.700673635" Nov 24 08:17:09 crc kubenswrapper[4799]: I1124 08:17:09.093336 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-wnwkw" podStartSLOduration=2.7782056 podStartE2EDuration="13.093310111s" podCreationTimestamp="2025-11-24 08:16:56 +0000 UTC" firstStartedPulling="2025-11-24 08:16:57.74098012 +0000 UTC m=+5363.396962584" lastFinishedPulling="2025-11-24 08:17:08.056084611 +0000 UTC m=+5373.712067095" observedRunningTime="2025-11-24 08:17:09.080965591 +0000 UTC m=+5374.736948085" watchObservedRunningTime="2025-11-24 08:17:09.093310111 +0000 UTC m=+5374.749292585" Nov 24 08:17:09 crc kubenswrapper[4799]: I1124 08:17:09.106554 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86db4855d-j9twl" podStartSLOduration=3.160330885 podStartE2EDuration="13.106534487s" podCreationTimestamp="2025-11-24 08:16:56 +0000 UTC" firstStartedPulling="2025-11-24 08:16:57.517363608 +0000 UTC m=+5363.173346082" lastFinishedPulling="2025-11-24 08:17:07.46356722 +0000 UTC m=+5373.119549684" observedRunningTime="2025-11-24 08:17:09.105472297 +0000 UTC m=+5374.761454771" watchObservedRunningTime="2025-11-24 08:17:09.106534487 +0000 UTC m=+5374.762516961" Nov 24 08:17:17 crc kubenswrapper[4799]: I1124 08:17:17.186491 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-26xdx" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.597116 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.597713 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="5e50e296-2811-4760-a259-86377d94888b" containerName="openstackclient" containerID="cri-o://aa8be354bf5825ffaa9fa768910a05d44954e65d34f59d8f186ecf39dce49c04" gracePeriod=2 Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.607047 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.671446 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 08:17:19 crc kubenswrapper[4799]: E1124 08:17:19.671946 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e50e296-2811-4760-a259-86377d94888b" containerName="openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.671965 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e50e296-2811-4760-a259-86377d94888b" containerName="openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.672161 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e50e296-2811-4760-a259-86377d94888b" containerName="openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.672941 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.686729 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.717294 4799 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c4bf26a-4262-47bf-ab27-9af023d43a8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:17:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.rdoproject.org/podified-antelope-centos9/openstack-openstackclient:94a34fda2d142cfe9e3097b1d1bd6839\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4plk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:17:19Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" not found" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.742188 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4plk\" (UniqueName: \"kubernetes.io/projected/7c4bf26a-4262-47bf-ab27-9af023d43a8f-kube-api-access-d4plk\") pod \"openstackclient\" (UID: \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\") " pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.742261 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config-secret\") pod \"openstackclient\" (UID: \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\") " pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.742307 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config\") pod \"openstackclient\" (UID: \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\") " pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.746475 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 08:17:19 crc kubenswrapper[4799]: E1124 08:17:19.784008 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-d4plk openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="7c4bf26a-4262-47bf-ab27-9af023d43a8f" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.784084 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.817173 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.818509 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.844007 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4plk\" (UniqueName: \"kubernetes.io/projected/7c4bf26a-4262-47bf-ab27-9af023d43a8f-kube-api-access-d4plk\") pod \"openstackclient\" (UID: \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\") " pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.844084 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config-secret\") pod \"openstackclient\" (UID: \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\") " pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.844136 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config\") pod \"openstackclient\" (UID: \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\") " pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.845327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config\") pod \"openstackclient\" (UID: \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\") " pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: E1124 08:17:19.851581 4799 projected.go:194] Error preparing data for projected volume kube-api-access-d4plk for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (7c4bf26a-4262-47bf-ab27-9af023d43a8f) does not match the UID in record. The object might have been deleted and then recreated Nov 24 08:17:19 crc kubenswrapper[4799]: E1124 08:17:19.851649 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7c4bf26a-4262-47bf-ab27-9af023d43a8f-kube-api-access-d4plk podName:7c4bf26a-4262-47bf-ab27-9af023d43a8f nodeName:}" failed. No retries permitted until 2025-11-24 08:17:20.351631232 +0000 UTC m=+5386.007613706 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-d4plk" (UniqueName: "kubernetes.io/projected/7c4bf26a-4262-47bf-ab27-9af023d43a8f-kube-api-access-d4plk") pod "openstackclient" (UID: "7c4bf26a-4262-47bf-ab27-9af023d43a8f") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (7c4bf26a-4262-47bf-ab27-9af023d43a8f) does not match the UID in record. The object might have been deleted and then recreated Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.877075 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.883646 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config-secret\") pod \"openstackclient\" (UID: \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\") " pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.891372 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="7c4bf26a-4262-47bf-ab27-9af023d43a8f" podUID="2093fc37-b94e-4a73-b661-d73b3f121c68" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.959412 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvf86\" (UniqueName: \"kubernetes.io/projected/2093fc37-b94e-4a73-b661-d73b3f121c68-kube-api-access-vvf86\") pod \"openstackclient\" (UID: \"2093fc37-b94e-4a73-b661-d73b3f121c68\") " pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.959543 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2093fc37-b94e-4a73-b661-d73b3f121c68-openstack-config\") pod \"openstackclient\" (UID: \"2093fc37-b94e-4a73-b661-d73b3f121c68\") " pod="openstack/openstackclient" Nov 24 08:17:19 crc kubenswrapper[4799]: I1124 08:17:19.959960 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2093fc37-b94e-4a73-b661-d73b3f121c68-openstack-config-secret\") pod \"openstackclient\" (UID: \"2093fc37-b94e-4a73-b661-d73b3f121c68\") " pod="openstack/openstackclient" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.006626 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.014288 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.019385 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-nmbfp" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.031163 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.061962 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2093fc37-b94e-4a73-b661-d73b3f121c68-openstack-config-secret\") pod \"openstackclient\" (UID: \"2093fc37-b94e-4a73-b661-d73b3f121c68\") " pod="openstack/openstackclient" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.062069 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdt5v\" (UniqueName: \"kubernetes.io/projected/051a8eb1-f406-4ab1-b59b-30754d729d0e-kube-api-access-gdt5v\") pod \"kube-state-metrics-0\" (UID: \"051a8eb1-f406-4ab1-b59b-30754d729d0e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.062109 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvf86\" (UniqueName: \"kubernetes.io/projected/2093fc37-b94e-4a73-b661-d73b3f121c68-kube-api-access-vvf86\") pod \"openstackclient\" (UID: \"2093fc37-b94e-4a73-b661-d73b3f121c68\") " pod="openstack/openstackclient" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.062143 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2093fc37-b94e-4a73-b661-d73b3f121c68-openstack-config\") pod \"openstackclient\" (UID: \"2093fc37-b94e-4a73-b661-d73b3f121c68\") " pod="openstack/openstackclient" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.062929 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2093fc37-b94e-4a73-b661-d73b3f121c68-openstack-config\") pod \"openstackclient\" (UID: \"2093fc37-b94e-4a73-b661-d73b3f121c68\") " pod="openstack/openstackclient" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.069465 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2093fc37-b94e-4a73-b661-d73b3f121c68-openstack-config-secret\") pod \"openstackclient\" (UID: \"2093fc37-b94e-4a73-b661-d73b3f121c68\") " pod="openstack/openstackclient" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.094737 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.100290 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="7c4bf26a-4262-47bf-ab27-9af023d43a8f" podUID="2093fc37-b94e-4a73-b661-d73b3f121c68" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.101372 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvf86\" (UniqueName: \"kubernetes.io/projected/2093fc37-b94e-4a73-b661-d73b3f121c68-kube-api-access-vvf86\") pod \"openstackclient\" (UID: \"2093fc37-b94e-4a73-b661-d73b3f121c68\") " pod="openstack/openstackclient" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.164598 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.165177 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdt5v\" (UniqueName: \"kubernetes.io/projected/051a8eb1-f406-4ab1-b59b-30754d729d0e-kube-api-access-gdt5v\") pod \"kube-state-metrics-0\" (UID: \"051a8eb1-f406-4ab1-b59b-30754d729d0e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.179993 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.187816 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="7c4bf26a-4262-47bf-ab27-9af023d43a8f" podUID="2093fc37-b94e-4a73-b661-d73b3f121c68" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.218420 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdt5v\" (UniqueName: \"kubernetes.io/projected/051a8eb1-f406-4ab1-b59b-30754d729d0e-kube-api-access-gdt5v\") pod \"kube-state-metrics-0\" (UID: \"051a8eb1-f406-4ab1-b59b-30754d729d0e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.276439 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config\") pod \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\" (UID: \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\") " Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.276571 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config-secret\") pod \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\" (UID: \"7c4bf26a-4262-47bf-ab27-9af023d43a8f\") " Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.277023 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4plk\" (UniqueName: \"kubernetes.io/projected/7c4bf26a-4262-47bf-ab27-9af023d43a8f-kube-api-access-d4plk\") on node \"crc\" DevicePath \"\"" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.277268 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "7c4bf26a-4262-47bf-ab27-9af023d43a8f" (UID: "7c4bf26a-4262-47bf-ab27-9af023d43a8f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.283931 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "7c4bf26a-4262-47bf-ab27-9af023d43a8f" (UID: "7c4bf26a-4262-47bf-ab27-9af023d43a8f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.345547 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.379039 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.379071 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c4bf26a-4262-47bf-ab27-9af023d43a8f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 08:17:20 crc kubenswrapper[4799]: W1124 08:17:20.869015 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2093fc37_b94e_4a73_b661_d73b3f121c68.slice/crio-4c70a05078eda06121ee2369bae721b8eca8acfae00c34c9140de03fd0d92ee7 WatchSource:0}: Error finding container 4c70a05078eda06121ee2369bae721b8eca8acfae00c34c9140de03fd0d92ee7: Status 404 returned error can't find the container with id 4c70a05078eda06121ee2369bae721b8eca8acfae00c34c9140de03fd0d92ee7 Nov 24 08:17:20 crc kubenswrapper[4799]: I1124 08:17:20.870495 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 08:17:21 crc kubenswrapper[4799]: W1124 08:17:21.055897 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod051a8eb1_f406_4ab1_b59b_30754d729d0e.slice/crio-9d6c2d4b79f770d7d4d7e9fca804964b54f992a5fdccc648802391682a26afef WatchSource:0}: Error finding container 9d6c2d4b79f770d7d4d7e9fca804964b54f992a5fdccc648802391682a26afef: Status 404 returned error can't find the container with id 9d6c2d4b79f770d7d4d7e9fca804964b54f992a5fdccc648802391682a26afef Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.061090 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.110994 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2093fc37-b94e-4a73-b661-d73b3f121c68","Type":"ContainerStarted","Data":"4c70a05078eda06121ee2369bae721b8eca8acfae00c34c9140de03fd0d92ee7"} Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.111925 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.113232 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"051a8eb1-f406-4ab1-b59b-30754d729d0e","Type":"ContainerStarted","Data":"9d6c2d4b79f770d7d4d7e9fca804964b54f992a5fdccc648802391682a26afef"} Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.115787 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="7c4bf26a-4262-47bf-ab27-9af023d43a8f" podUID="2093fc37-b94e-4a73-b661-d73b3f121c68" Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.136810 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="7c4bf26a-4262-47bf-ab27-9af023d43a8f" podUID="2093fc37-b94e-4a73-b661-d73b3f121c68" Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.686571 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c4bf26a-4262-47bf-ab27-9af023d43a8f" path="/var/lib/kubelet/pods/7c4bf26a-4262-47bf-ab27-9af023d43a8f/volumes" Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.918433 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.920364 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.922178 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.922239 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-mgcdg" Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.923184 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.924003 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.924180 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Nov 24 08:17:21 crc kubenswrapper[4799]: I1124 08:17:21.947177 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.018906 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.018970 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vnrh\" (UniqueName: \"kubernetes.io/projected/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-kube-api-access-8vnrh\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.019028 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.019058 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.019085 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.019137 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.019163 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.121453 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.121883 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.121987 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vnrh\" (UniqueName: \"kubernetes.io/projected/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-kube-api-access-8vnrh\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.122387 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.122420 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.122444 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.122500 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.122522 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.124106 4799 generic.go:334] "Generic (PLEG): container finished" podID="5e50e296-2811-4760-a259-86377d94888b" containerID="aa8be354bf5825ffaa9fa768910a05d44954e65d34f59d8f186ecf39dce49c04" exitCode=137 Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.128705 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.131683 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.139720 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.142142 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.144522 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"051a8eb1-f406-4ab1-b59b-30754d729d0e","Type":"ContainerStarted","Data":"f112ce21517751d136fff04e1503601ea2bd576aa76d4e94ef201ec3938c2931"} Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.145491 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vnrh\" (UniqueName: \"kubernetes.io/projected/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-kube-api-access-8vnrh\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.145884 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.146123 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fc77dc6a-33f2-4d03-a7f8-1699e2d37fab-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.159410 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2093fc37-b94e-4a73-b661-d73b3f121c68","Type":"ContainerStarted","Data":"f905f3623c64794e45246c7b6a1f24c77c9d78c0a961f50c09ff58a27aa2e48c"} Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.182130 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.80653458 podStartE2EDuration="3.182094507s" podCreationTimestamp="2025-11-24 08:17:19 +0000 UTC" firstStartedPulling="2025-11-24 08:17:21.060484829 +0000 UTC m=+5386.716467303" lastFinishedPulling="2025-11-24 08:17:21.436044756 +0000 UTC m=+5387.092027230" observedRunningTime="2025-11-24 08:17:22.179404751 +0000 UTC m=+5387.835387225" watchObservedRunningTime="2025-11-24 08:17:22.182094507 +0000 UTC m=+5387.838076981" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.217365 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.217343118 podStartE2EDuration="3.217343118s" podCreationTimestamp="2025-11-24 08:17:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:22.200722426 +0000 UTC m=+5387.856704900" watchObservedRunningTime="2025-11-24 08:17:22.217343118 +0000 UTC m=+5387.873325592" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.291913 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.437868 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.440251 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.443513 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.443577 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.443720 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.443927 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.449072 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-grbtn" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.455248 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.461330 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.536896 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-config\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.536936 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.536973 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.537027 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.537053 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.537073 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.537113 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-61ee0d73-4614-4c9e-8f7b-503de2bc54dd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61ee0d73-4614-4c9e-8f7b-503de2bc54dd\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.537138 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv69n\" (UniqueName: \"kubernetes.io/projected/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-kube-api-access-tv69n\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.638181 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-config\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.638231 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.638280 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.638334 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.638362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.638381 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.638423 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-61ee0d73-4614-4c9e-8f7b-503de2bc54dd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61ee0d73-4614-4c9e-8f7b-503de2bc54dd\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.638449 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv69n\" (UniqueName: \"kubernetes.io/projected/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-kube-api-access-tv69n\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.640911 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.649625 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.650273 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.652781 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.652802 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.662506 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-config\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.663379 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.686667 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.686710 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-61ee0d73-4614-4c9e-8f7b-503de2bc54dd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61ee0d73-4614-4c9e-8f7b-503de2bc54dd\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5cb82ff7e01639fa11cc6583bf853ebc48269582932b00ea5b1a805353e90c98/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.690582 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv69n\" (UniqueName: \"kubernetes.io/projected/d2f7efbf-fb81-4cb1-8284-42baf4c8da5e-kube-api-access-tv69n\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.727817 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-61ee0d73-4614-4c9e-8f7b-503de2bc54dd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-61ee0d73-4614-4c9e-8f7b-503de2bc54dd\") pod \"prometheus-metric-storage-0\" (UID: \"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e\") " pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.739363 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5e50e296-2811-4760-a259-86377d94888b-openstack-config-secret\") pod \"5e50e296-2811-4760-a259-86377d94888b\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.739663 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5e50e296-2811-4760-a259-86377d94888b-openstack-config\") pod \"5e50e296-2811-4760-a259-86377d94888b\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.739866 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bw6m\" (UniqueName: \"kubernetes.io/projected/5e50e296-2811-4760-a259-86377d94888b-kube-api-access-8bw6m\") pod \"5e50e296-2811-4760-a259-86377d94888b\" (UID: \"5e50e296-2811-4760-a259-86377d94888b\") " Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.743214 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e50e296-2811-4760-a259-86377d94888b-kube-api-access-8bw6m" (OuterVolumeSpecName: "kube-api-access-8bw6m") pod "5e50e296-2811-4760-a259-86377d94888b" (UID: "5e50e296-2811-4760-a259-86377d94888b"). InnerVolumeSpecName "kube-api-access-8bw6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.766429 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e50e296-2811-4760-a259-86377d94888b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5e50e296-2811-4760-a259-86377d94888b" (UID: "5e50e296-2811-4760-a259-86377d94888b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.769896 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.798978 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e50e296-2811-4760-a259-86377d94888b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5e50e296-2811-4760-a259-86377d94888b" (UID: "5e50e296-2811-4760-a259-86377d94888b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.830115 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 08:17:22 crc kubenswrapper[4799]: W1124 08:17:22.833575 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc77dc6a_33f2_4d03_a7f8_1699e2d37fab.slice/crio-23b227e07fd40e8426950ece4705982df346cefa7e65422de5196a7873cc4e49 WatchSource:0}: Error finding container 23b227e07fd40e8426950ece4705982df346cefa7e65422de5196a7873cc4e49: Status 404 returned error can't find the container with id 23b227e07fd40e8426950ece4705982df346cefa7e65422de5196a7873cc4e49 Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.849117 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bw6m\" (UniqueName: \"kubernetes.io/projected/5e50e296-2811-4760-a259-86377d94888b-kube-api-access-8bw6m\") on node \"crc\" DevicePath \"\"" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.849157 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5e50e296-2811-4760-a259-86377d94888b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 08:17:22 crc kubenswrapper[4799]: I1124 08:17:22.849170 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5e50e296-2811-4760-a259-86377d94888b-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:17:23 crc kubenswrapper[4799]: I1124 08:17:23.170007 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab","Type":"ContainerStarted","Data":"23b227e07fd40e8426950ece4705982df346cefa7e65422de5196a7873cc4e49"} Nov 24 08:17:23 crc kubenswrapper[4799]: I1124 08:17:23.173925 4799 scope.go:117] "RemoveContainer" containerID="aa8be354bf5825ffaa9fa768910a05d44954e65d34f59d8f186ecf39dce49c04" Nov 24 08:17:23 crc kubenswrapper[4799]: I1124 08:17:23.174007 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:17:23 crc kubenswrapper[4799]: I1124 08:17:23.193292 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5e50e296-2811-4760-a259-86377d94888b" podUID="2093fc37-b94e-4a73-b661-d73b3f121c68" Nov 24 08:17:23 crc kubenswrapper[4799]: E1124 08:17:23.250824 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e50e296_2811_4760_a259_86377d94888b.slice/crio-022bed948c3e937661b18a27aa78e54fe490ac3940d0fa1f427a30f8260feaaa\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e50e296_2811_4760_a259_86377d94888b.slice\": RecentStats: unable to find data in memory cache]" Nov 24 08:17:23 crc kubenswrapper[4799]: I1124 08:17:23.286590 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 08:17:23 crc kubenswrapper[4799]: W1124 08:17:23.290460 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2f7efbf_fb81_4cb1_8284_42baf4c8da5e.slice/crio-637366fe73dc44bf6be9ed90762d0f2bfbabcac80e8f27181b0ff54fbe9770e2 WatchSource:0}: Error finding container 637366fe73dc44bf6be9ed90762d0f2bfbabcac80e8f27181b0ff54fbe9770e2: Status 404 returned error can't find the container with id 637366fe73dc44bf6be9ed90762d0f2bfbabcac80e8f27181b0ff54fbe9770e2 Nov 24 08:17:23 crc kubenswrapper[4799]: I1124 08:17:23.641087 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e50e296-2811-4760-a259-86377d94888b" path="/var/lib/kubelet/pods/5e50e296-2811-4760-a259-86377d94888b/volumes" Nov 24 08:17:24 crc kubenswrapper[4799]: I1124 08:17:24.184953 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e","Type":"ContainerStarted","Data":"637366fe73dc44bf6be9ed90762d0f2bfbabcac80e8f27181b0ff54fbe9770e2"} Nov 24 08:17:29 crc kubenswrapper[4799]: I1124 08:17:29.275679 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab","Type":"ContainerStarted","Data":"bf96b4c27d38dc651e30bb3b0c2d49decc99a7aa701a8b1627ddbff359374a94"} Nov 24 08:17:30 crc kubenswrapper[4799]: I1124 08:17:30.288047 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e","Type":"ContainerStarted","Data":"bb31240c9974e733ec1d4e0d1bb01a55b68b115d74dbb341c4f2c6af22547f8a"} Nov 24 08:17:30 crc kubenswrapper[4799]: I1124 08:17:30.349290 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 08:17:38 crc kubenswrapper[4799]: I1124 08:17:38.376801 4799 generic.go:334] "Generic (PLEG): container finished" podID="fc77dc6a-33f2-4d03-a7f8-1699e2d37fab" containerID="bf96b4c27d38dc651e30bb3b0c2d49decc99a7aa701a8b1627ddbff359374a94" exitCode=0 Nov 24 08:17:38 crc kubenswrapper[4799]: I1124 08:17:38.376912 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab","Type":"ContainerDied","Data":"bf96b4c27d38dc651e30bb3b0c2d49decc99a7aa701a8b1627ddbff359374a94"} Nov 24 08:17:38 crc kubenswrapper[4799]: I1124 08:17:38.384981 4799 generic.go:334] "Generic (PLEG): container finished" podID="d2f7efbf-fb81-4cb1-8284-42baf4c8da5e" containerID="bb31240c9974e733ec1d4e0d1bb01a55b68b115d74dbb341c4f2c6af22547f8a" exitCode=0 Nov 24 08:17:38 crc kubenswrapper[4799]: I1124 08:17:38.385034 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e","Type":"ContainerDied","Data":"bb31240c9974e733ec1d4e0d1bb01a55b68b115d74dbb341c4f2c6af22547f8a"} Nov 24 08:17:41 crc kubenswrapper[4799]: I1124 08:17:41.047704 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-lcv54"] Nov 24 08:17:41 crc kubenswrapper[4799]: I1124 08:17:41.058430 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-64b3-account-create-vrs9s"] Nov 24 08:17:41 crc kubenswrapper[4799]: I1124 08:17:41.068712 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-64b3-account-create-vrs9s"] Nov 24 08:17:41 crc kubenswrapper[4799]: I1124 08:17:41.077674 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-lcv54"] Nov 24 08:17:41 crc kubenswrapper[4799]: I1124 08:17:41.641568 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1613cebb-f738-4ad3-bdcb-9d9553c0505a" path="/var/lib/kubelet/pods/1613cebb-f738-4ad3-bdcb-9d9553c0505a/volumes" Nov 24 08:17:41 crc kubenswrapper[4799]: I1124 08:17:41.642887 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35" path="/var/lib/kubelet/pods/3bea8cc5-3dd3-4719-8b6e-cdb66baa4a35/volumes" Nov 24 08:17:42 crc kubenswrapper[4799]: I1124 08:17:42.573446 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab","Type":"ContainerStarted","Data":"bd72e8d27f64a4ae2ec9668ff4d210f28b439740c4ad974bfdfb375c3f722cce"} Nov 24 08:17:45 crc kubenswrapper[4799]: I1124 08:17:45.609775 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"fc77dc6a-33f2-4d03-a7f8-1699e2d37fab","Type":"ContainerStarted","Data":"14d5afaedab3113b9680bf931212d45ecc93feb64b5037cf48af5eeb6781e9f5"} Nov 24 08:17:45 crc kubenswrapper[4799]: I1124 08:17:45.610333 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:45 crc kubenswrapper[4799]: I1124 08:17:45.611980 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e","Type":"ContainerStarted","Data":"834f97d235622b68c071ec49e0932736a981764d4c00d95057f5b88c35dfeccb"} Nov 24 08:17:45 crc kubenswrapper[4799]: I1124 08:17:45.612109 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Nov 24 08:17:45 crc kubenswrapper[4799]: I1124 08:17:45.649872 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=5.919754106 podStartE2EDuration="24.649819895s" podCreationTimestamp="2025-11-24 08:17:21 +0000 UTC" firstStartedPulling="2025-11-24 08:17:22.836055282 +0000 UTC m=+5388.492037756" lastFinishedPulling="2025-11-24 08:17:41.566121061 +0000 UTC m=+5407.222103545" observedRunningTime="2025-11-24 08:17:45.64227091 +0000 UTC m=+5411.298253384" watchObservedRunningTime="2025-11-24 08:17:45.649819895 +0000 UTC m=+5411.305802389" Nov 24 08:17:49 crc kubenswrapper[4799]: I1124 08:17:49.660229 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e","Type":"ContainerStarted","Data":"749501692abb69e87a6ff6db2c00755317c1697888b9d2ea1536634fc6bf9152"} Nov 24 08:17:50 crc kubenswrapper[4799]: I1124 08:17:50.401258 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:17:50 crc kubenswrapper[4799]: I1124 08:17:50.401331 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:17:53 crc kubenswrapper[4799]: I1124 08:17:53.732320 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d2f7efbf-fb81-4cb1-8284-42baf4c8da5e","Type":"ContainerStarted","Data":"e58f94092ec93f4290bb57226d53b4ba152fdfd682402abc0b4d007f9fde9750"} Nov 24 08:17:53 crc kubenswrapper[4799]: I1124 08:17:53.821512 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.581179122 podStartE2EDuration="32.821486237s" podCreationTimestamp="2025-11-24 08:17:21 +0000 UTC" firstStartedPulling="2025-11-24 08:17:23.292527498 +0000 UTC m=+5388.948509972" lastFinishedPulling="2025-11-24 08:17:52.532834613 +0000 UTC m=+5418.188817087" observedRunningTime="2025-11-24 08:17:53.805530003 +0000 UTC m=+5419.461512477" watchObservedRunningTime="2025-11-24 08:17:53.821486237 +0000 UTC m=+5419.477468751" Nov 24 08:17:57 crc kubenswrapper[4799]: I1124 08:17:57.770108 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.720137 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.723646 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.729115 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.733628 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.740354 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.883545 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.883595 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-run-httpd\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.883618 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.883697 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-log-httpd\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.883748 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-config-data\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.883771 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vl2k\" (UniqueName: \"kubernetes.io/projected/5ea732c6-c837-4ea5-ad1f-d226502ef827-kube-api-access-9vl2k\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.883786 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-scripts\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.985232 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.985299 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-run-httpd\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.985331 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.985504 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-log-httpd\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.985567 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-config-data\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.985597 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vl2k\" (UniqueName: \"kubernetes.io/projected/5ea732c6-c837-4ea5-ad1f-d226502ef827-kube-api-access-9vl2k\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.985620 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-scripts\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.985901 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-run-httpd\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.986215 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-log-httpd\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.992332 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.992337 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.993751 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-config-data\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:58 crc kubenswrapper[4799]: I1124 08:17:58.994779 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-scripts\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:59 crc kubenswrapper[4799]: I1124 08:17:59.012061 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vl2k\" (UniqueName: \"kubernetes.io/projected/5ea732c6-c837-4ea5-ad1f-d226502ef827-kube-api-access-9vl2k\") pod \"ceilometer-0\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " pod="openstack/ceilometer-0" Nov 24 08:17:59 crc kubenswrapper[4799]: I1124 08:17:59.051059 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:17:59 crc kubenswrapper[4799]: I1124 08:17:59.593459 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:17:59 crc kubenswrapper[4799]: I1124 08:17:59.824132 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ea732c6-c837-4ea5-ad1f-d226502ef827","Type":"ContainerStarted","Data":"540fe33108ec2671488f930ebbe4b57f9f1e1fbbebce5d180c40cb1d1cab5b56"} Nov 24 08:18:04 crc kubenswrapper[4799]: I1124 08:18:04.877194 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ea732c6-c837-4ea5-ad1f-d226502ef827","Type":"ContainerStarted","Data":"be39a2e3da5f849c1b00c4a2e3dc0df0c676e2a1fdb73c19cf1342702a847e81"} Nov 24 08:18:05 crc kubenswrapper[4799]: I1124 08:18:05.890674 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ea732c6-c837-4ea5-ad1f-d226502ef827","Type":"ContainerStarted","Data":"4bbdb6b5015e0bfb5b2b32149f27e147b65b647f6917662a2f142775cd5a97d7"} Nov 24 08:18:06 crc kubenswrapper[4799]: I1124 08:18:06.070935 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-jl8c5"] Nov 24 08:18:06 crc kubenswrapper[4799]: I1124 08:18:06.081394 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-jl8c5"] Nov 24 08:18:06 crc kubenswrapper[4799]: I1124 08:18:06.902461 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ea732c6-c837-4ea5-ad1f-d226502ef827","Type":"ContainerStarted","Data":"edf175a59ce367dadd20f6a930891b369beb7b84d9e933e52460e5ddf4adea46"} Nov 24 08:18:07 crc kubenswrapper[4799]: I1124 08:18:07.646764 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f99a056-2251-45ca-a55e-f970ab0fe1e2" path="/var/lib/kubelet/pods/8f99a056-2251-45ca-a55e-f970ab0fe1e2/volumes" Nov 24 08:18:07 crc kubenswrapper[4799]: I1124 08:18:07.771365 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 24 08:18:07 crc kubenswrapper[4799]: I1124 08:18:07.783286 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 24 08:18:07 crc kubenswrapper[4799]: I1124 08:18:07.919134 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 24 08:18:08 crc kubenswrapper[4799]: I1124 08:18:08.324647 4799 scope.go:117] "RemoveContainer" containerID="f3e4727f4ba08177b8461aafc80be8621d5bf160338f88797cbb79336e866810" Nov 24 08:18:08 crc kubenswrapper[4799]: I1124 08:18:08.370301 4799 scope.go:117] "RemoveContainer" containerID="1dc3fc380db55fd127a6133a8928eb1cdcc9be66e2dbe95072bd3574e2cc5235" Nov 24 08:18:08 crc kubenswrapper[4799]: I1124 08:18:08.401952 4799 scope.go:117] "RemoveContainer" containerID="d7e327782ee7420167fc4fa23e5b79e12160ed533afde1fe1838da935cc9f6cb" Nov 24 08:18:08 crc kubenswrapper[4799]: I1124 08:18:08.928435 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ea732c6-c837-4ea5-ad1f-d226502ef827","Type":"ContainerStarted","Data":"37a9cbbdd7ca4e0e9d83d34193a8ea0a846f64002da5d9826ccfb0761e1a9abb"} Nov 24 08:18:08 crc kubenswrapper[4799]: I1124 08:18:08.928504 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 08:18:08 crc kubenswrapper[4799]: I1124 08:18:08.959163 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.677777113 podStartE2EDuration="10.959142777s" podCreationTimestamp="2025-11-24 08:17:58 +0000 UTC" firstStartedPulling="2025-11-24 08:17:59.599286749 +0000 UTC m=+5425.255269223" lastFinishedPulling="2025-11-24 08:18:07.880652413 +0000 UTC m=+5433.536634887" observedRunningTime="2025-11-24 08:18:08.947381233 +0000 UTC m=+5434.603363707" watchObservedRunningTime="2025-11-24 08:18:08.959142777 +0000 UTC m=+5434.615125251" Nov 24 08:18:15 crc kubenswrapper[4799]: I1124 08:18:15.818643 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-w4dc6"] Nov 24 08:18:15 crc kubenswrapper[4799]: I1124 08:18:15.820557 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-w4dc6" Nov 24 08:18:15 crc kubenswrapper[4799]: I1124 08:18:15.828383 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-w4dc6"] Nov 24 08:18:15 crc kubenswrapper[4799]: I1124 08:18:15.913080 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d87dc8df-3799-495f-8090-536332563977-operator-scripts\") pod \"aodh-db-create-w4dc6\" (UID: \"d87dc8df-3799-495f-8090-536332563977\") " pod="openstack/aodh-db-create-w4dc6" Nov 24 08:18:15 crc kubenswrapper[4799]: I1124 08:18:15.913163 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7jdn\" (UniqueName: \"kubernetes.io/projected/d87dc8df-3799-495f-8090-536332563977-kube-api-access-f7jdn\") pod \"aodh-db-create-w4dc6\" (UID: \"d87dc8df-3799-495f-8090-536332563977\") " pod="openstack/aodh-db-create-w4dc6" Nov 24 08:18:15 crc kubenswrapper[4799]: I1124 08:18:15.913211 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-ec59-account-create-8nmkj"] Nov 24 08:18:15 crc kubenswrapper[4799]: I1124 08:18:15.914726 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-ec59-account-create-8nmkj" Nov 24 08:18:15 crc kubenswrapper[4799]: I1124 08:18:15.919316 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Nov 24 08:18:15 crc kubenswrapper[4799]: I1124 08:18:15.923280 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-ec59-account-create-8nmkj"] Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.014879 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d87dc8df-3799-495f-8090-536332563977-operator-scripts\") pod \"aodh-db-create-w4dc6\" (UID: \"d87dc8df-3799-495f-8090-536332563977\") " pod="openstack/aodh-db-create-w4dc6" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.014986 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7jdn\" (UniqueName: \"kubernetes.io/projected/d87dc8df-3799-495f-8090-536332563977-kube-api-access-f7jdn\") pod \"aodh-db-create-w4dc6\" (UID: \"d87dc8df-3799-495f-8090-536332563977\") " pod="openstack/aodh-db-create-w4dc6" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.015163 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31932a1e-5342-4097-8e42-8855cf672708-operator-scripts\") pod \"aodh-ec59-account-create-8nmkj\" (UID: \"31932a1e-5342-4097-8e42-8855cf672708\") " pod="openstack/aodh-ec59-account-create-8nmkj" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.015193 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk94k\" (UniqueName: \"kubernetes.io/projected/31932a1e-5342-4097-8e42-8855cf672708-kube-api-access-tk94k\") pod \"aodh-ec59-account-create-8nmkj\" (UID: \"31932a1e-5342-4097-8e42-8855cf672708\") " pod="openstack/aodh-ec59-account-create-8nmkj" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.015816 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d87dc8df-3799-495f-8090-536332563977-operator-scripts\") pod \"aodh-db-create-w4dc6\" (UID: \"d87dc8df-3799-495f-8090-536332563977\") " pod="openstack/aodh-db-create-w4dc6" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.034146 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7jdn\" (UniqueName: \"kubernetes.io/projected/d87dc8df-3799-495f-8090-536332563977-kube-api-access-f7jdn\") pod \"aodh-db-create-w4dc6\" (UID: \"d87dc8df-3799-495f-8090-536332563977\") " pod="openstack/aodh-db-create-w4dc6" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.117386 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31932a1e-5342-4097-8e42-8855cf672708-operator-scripts\") pod \"aodh-ec59-account-create-8nmkj\" (UID: \"31932a1e-5342-4097-8e42-8855cf672708\") " pod="openstack/aodh-ec59-account-create-8nmkj" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.117441 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk94k\" (UniqueName: \"kubernetes.io/projected/31932a1e-5342-4097-8e42-8855cf672708-kube-api-access-tk94k\") pod \"aodh-ec59-account-create-8nmkj\" (UID: \"31932a1e-5342-4097-8e42-8855cf672708\") " pod="openstack/aodh-ec59-account-create-8nmkj" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.118510 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31932a1e-5342-4097-8e42-8855cf672708-operator-scripts\") pod \"aodh-ec59-account-create-8nmkj\" (UID: \"31932a1e-5342-4097-8e42-8855cf672708\") " pod="openstack/aodh-ec59-account-create-8nmkj" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.137048 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk94k\" (UniqueName: \"kubernetes.io/projected/31932a1e-5342-4097-8e42-8855cf672708-kube-api-access-tk94k\") pod \"aodh-ec59-account-create-8nmkj\" (UID: \"31932a1e-5342-4097-8e42-8855cf672708\") " pod="openstack/aodh-ec59-account-create-8nmkj" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.151782 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-w4dc6" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.237954 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-ec59-account-create-8nmkj" Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.691320 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-w4dc6"] Nov 24 08:18:16 crc kubenswrapper[4799]: W1124 08:18:16.693485 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd87dc8df_3799_495f_8090_536332563977.slice/crio-30be29d08e9c0a1eb4fea4b4ac9950adf927620d4cc36e17a61df68710981ff2 WatchSource:0}: Error finding container 30be29d08e9c0a1eb4fea4b4ac9950adf927620d4cc36e17a61df68710981ff2: Status 404 returned error can't find the container with id 30be29d08e9c0a1eb4fea4b4ac9950adf927620d4cc36e17a61df68710981ff2 Nov 24 08:18:16 crc kubenswrapper[4799]: I1124 08:18:16.808691 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-ec59-account-create-8nmkj"] Nov 24 08:18:16 crc kubenswrapper[4799]: W1124 08:18:16.816243 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31932a1e_5342_4097_8e42_8855cf672708.slice/crio-e7eab4c770872226935e399971cecad6f544bd93a634d2ef71f822f80a99af6f WatchSource:0}: Error finding container e7eab4c770872226935e399971cecad6f544bd93a634d2ef71f822f80a99af6f: Status 404 returned error can't find the container with id e7eab4c770872226935e399971cecad6f544bd93a634d2ef71f822f80a99af6f Nov 24 08:18:17 crc kubenswrapper[4799]: I1124 08:18:17.013128 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-w4dc6" event={"ID":"d87dc8df-3799-495f-8090-536332563977","Type":"ContainerStarted","Data":"6db2b871502bd4c16168a3adc2db8d87a64ab75a32d7901a1135f05dbe9bf1fd"} Nov 24 08:18:17 crc kubenswrapper[4799]: I1124 08:18:17.013172 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-w4dc6" event={"ID":"d87dc8df-3799-495f-8090-536332563977","Type":"ContainerStarted","Data":"30be29d08e9c0a1eb4fea4b4ac9950adf927620d4cc36e17a61df68710981ff2"} Nov 24 08:18:17 crc kubenswrapper[4799]: I1124 08:18:17.015843 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-ec59-account-create-8nmkj" event={"ID":"31932a1e-5342-4097-8e42-8855cf672708","Type":"ContainerStarted","Data":"562e27c5270effa8f4a9ab891eb34f5ca34b1515e27a9117556cc4ea7dc0e67e"} Nov 24 08:18:17 crc kubenswrapper[4799]: I1124 08:18:17.015916 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-ec59-account-create-8nmkj" event={"ID":"31932a1e-5342-4097-8e42-8855cf672708","Type":"ContainerStarted","Data":"e7eab4c770872226935e399971cecad6f544bd93a634d2ef71f822f80a99af6f"} Nov 24 08:18:17 crc kubenswrapper[4799]: I1124 08:18:17.035496 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-w4dc6" podStartSLOduration=2.035467938 podStartE2EDuration="2.035467938s" podCreationTimestamp="2025-11-24 08:18:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:18:17.0264038 +0000 UTC m=+5442.682386264" watchObservedRunningTime="2025-11-24 08:18:17.035467938 +0000 UTC m=+5442.691450412" Nov 24 08:18:17 crc kubenswrapper[4799]: I1124 08:18:17.069758 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-ec59-account-create-8nmkj" podStartSLOduration=2.069731061 podStartE2EDuration="2.069731061s" podCreationTimestamp="2025-11-24 08:18:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:18:17.06477797 +0000 UTC m=+5442.720760454" watchObservedRunningTime="2025-11-24 08:18:17.069731061 +0000 UTC m=+5442.725713535" Nov 24 08:18:18 crc kubenswrapper[4799]: I1124 08:18:18.027243 4799 generic.go:334] "Generic (PLEG): container finished" podID="31932a1e-5342-4097-8e42-8855cf672708" containerID="562e27c5270effa8f4a9ab891eb34f5ca34b1515e27a9117556cc4ea7dc0e67e" exitCode=0 Nov 24 08:18:18 crc kubenswrapper[4799]: I1124 08:18:18.028022 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-ec59-account-create-8nmkj" event={"ID":"31932a1e-5342-4097-8e42-8855cf672708","Type":"ContainerDied","Data":"562e27c5270effa8f4a9ab891eb34f5ca34b1515e27a9117556cc4ea7dc0e67e"} Nov 24 08:18:18 crc kubenswrapper[4799]: I1124 08:18:18.029583 4799 generic.go:334] "Generic (PLEG): container finished" podID="d87dc8df-3799-495f-8090-536332563977" containerID="6db2b871502bd4c16168a3adc2db8d87a64ab75a32d7901a1135f05dbe9bf1fd" exitCode=0 Nov 24 08:18:18 crc kubenswrapper[4799]: I1124 08:18:18.029652 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-w4dc6" event={"ID":"d87dc8df-3799-495f-8090-536332563977","Type":"ContainerDied","Data":"6db2b871502bd4c16168a3adc2db8d87a64ab75a32d7901a1135f05dbe9bf1fd"} Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.563355 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-w4dc6" Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.569965 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-ec59-account-create-8nmkj" Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.688924 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31932a1e-5342-4097-8e42-8855cf672708-operator-scripts\") pod \"31932a1e-5342-4097-8e42-8855cf672708\" (UID: \"31932a1e-5342-4097-8e42-8855cf672708\") " Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.689225 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7jdn\" (UniqueName: \"kubernetes.io/projected/d87dc8df-3799-495f-8090-536332563977-kube-api-access-f7jdn\") pod \"d87dc8df-3799-495f-8090-536332563977\" (UID: \"d87dc8df-3799-495f-8090-536332563977\") " Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.689406 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk94k\" (UniqueName: \"kubernetes.io/projected/31932a1e-5342-4097-8e42-8855cf672708-kube-api-access-tk94k\") pod \"31932a1e-5342-4097-8e42-8855cf672708\" (UID: \"31932a1e-5342-4097-8e42-8855cf672708\") " Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.689473 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d87dc8df-3799-495f-8090-536332563977-operator-scripts\") pod \"d87dc8df-3799-495f-8090-536332563977\" (UID: \"d87dc8df-3799-495f-8090-536332563977\") " Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.689876 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31932a1e-5342-4097-8e42-8855cf672708-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "31932a1e-5342-4097-8e42-8855cf672708" (UID: "31932a1e-5342-4097-8e42-8855cf672708"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.689908 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d87dc8df-3799-495f-8090-536332563977-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d87dc8df-3799-495f-8090-536332563977" (UID: "d87dc8df-3799-495f-8090-536332563977"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.694335 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d87dc8df-3799-495f-8090-536332563977-kube-api-access-f7jdn" (OuterVolumeSpecName: "kube-api-access-f7jdn") pod "d87dc8df-3799-495f-8090-536332563977" (UID: "d87dc8df-3799-495f-8090-536332563977"). InnerVolumeSpecName "kube-api-access-f7jdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.696386 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31932a1e-5342-4097-8e42-8855cf672708-kube-api-access-tk94k" (OuterVolumeSpecName: "kube-api-access-tk94k") pod "31932a1e-5342-4097-8e42-8855cf672708" (UID: "31932a1e-5342-4097-8e42-8855cf672708"). InnerVolumeSpecName "kube-api-access-tk94k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.791262 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk94k\" (UniqueName: \"kubernetes.io/projected/31932a1e-5342-4097-8e42-8855cf672708-kube-api-access-tk94k\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.791288 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d87dc8df-3799-495f-8090-536332563977-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.791297 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31932a1e-5342-4097-8e42-8855cf672708-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:19 crc kubenswrapper[4799]: I1124 08:18:19.791305 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7jdn\" (UniqueName: \"kubernetes.io/projected/d87dc8df-3799-495f-8090-536332563977-kube-api-access-f7jdn\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:20 crc kubenswrapper[4799]: I1124 08:18:20.057345 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-w4dc6" event={"ID":"d87dc8df-3799-495f-8090-536332563977","Type":"ContainerDied","Data":"30be29d08e9c0a1eb4fea4b4ac9950adf927620d4cc36e17a61df68710981ff2"} Nov 24 08:18:20 crc kubenswrapper[4799]: I1124 08:18:20.057401 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-w4dc6" Nov 24 08:18:20 crc kubenswrapper[4799]: I1124 08:18:20.057408 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30be29d08e9c0a1eb4fea4b4ac9950adf927620d4cc36e17a61df68710981ff2" Nov 24 08:18:20 crc kubenswrapper[4799]: I1124 08:18:20.059575 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-ec59-account-create-8nmkj" event={"ID":"31932a1e-5342-4097-8e42-8855cf672708","Type":"ContainerDied","Data":"e7eab4c770872226935e399971cecad6f544bd93a634d2ef71f822f80a99af6f"} Nov 24 08:18:20 crc kubenswrapper[4799]: I1124 08:18:20.059622 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7eab4c770872226935e399971cecad6f544bd93a634d2ef71f822f80a99af6f" Nov 24 08:18:20 crc kubenswrapper[4799]: I1124 08:18:20.059657 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-ec59-account-create-8nmkj" Nov 24 08:18:20 crc kubenswrapper[4799]: I1124 08:18:20.400685 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:18:20 crc kubenswrapper[4799]: I1124 08:18:20.400746 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.189882 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-wvq24"] Nov 24 08:18:21 crc kubenswrapper[4799]: E1124 08:18:21.190656 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31932a1e-5342-4097-8e42-8855cf672708" containerName="mariadb-account-create" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.190673 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="31932a1e-5342-4097-8e42-8855cf672708" containerName="mariadb-account-create" Nov 24 08:18:21 crc kubenswrapper[4799]: E1124 08:18:21.190712 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d87dc8df-3799-495f-8090-536332563977" containerName="mariadb-database-create" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.190717 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d87dc8df-3799-495f-8090-536332563977" containerName="mariadb-database-create" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.190930 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="31932a1e-5342-4097-8e42-8855cf672708" containerName="mariadb-account-create" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.190959 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d87dc8df-3799-495f-8090-536332563977" containerName="mariadb-database-create" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.191663 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.194317 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.194547 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.194693 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-qlvql" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.195168 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.220597 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-wvq24"] Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.323516 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-scripts\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.323795 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-combined-ca-bundle\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.323998 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p9b9\" (UniqueName: \"kubernetes.io/projected/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-kube-api-access-4p9b9\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.324160 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-config-data\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.426292 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-scripts\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.426658 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-combined-ca-bundle\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.426960 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p9b9\" (UniqueName: \"kubernetes.io/projected/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-kube-api-access-4p9b9\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.427263 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-config-data\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.433081 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-scripts\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.433281 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-combined-ca-bundle\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.435060 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-config-data\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.445229 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p9b9\" (UniqueName: \"kubernetes.io/projected/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-kube-api-access-4p9b9\") pod \"aodh-db-sync-wvq24\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:21 crc kubenswrapper[4799]: I1124 08:18:21.514610 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:22 crc kubenswrapper[4799]: I1124 08:18:22.019021 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-wvq24"] Nov 24 08:18:22 crc kubenswrapper[4799]: I1124 08:18:22.076799 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wvq24" event={"ID":"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81","Type":"ContainerStarted","Data":"e4ad8d9ed00054196a936be5a6d2f794814a15d776c2f264c1806199adb248bc"} Nov 24 08:18:28 crc kubenswrapper[4799]: I1124 08:18:28.166923 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wvq24" event={"ID":"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81","Type":"ContainerStarted","Data":"dce6e19b87b53b1c51b330203a18e4aa042fc14397acf2e0e92ffe3d992713a1"} Nov 24 08:18:28 crc kubenswrapper[4799]: I1124 08:18:28.185122 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-wvq24" podStartSLOduration=2.115214826 podStartE2EDuration="7.185101712s" podCreationTimestamp="2025-11-24 08:18:21 +0000 UTC" firstStartedPulling="2025-11-24 08:18:22.028554441 +0000 UTC m=+5447.684536905" lastFinishedPulling="2025-11-24 08:18:27.098441317 +0000 UTC m=+5452.754423791" observedRunningTime="2025-11-24 08:18:28.184396752 +0000 UTC m=+5453.840379226" watchObservedRunningTime="2025-11-24 08:18:28.185101712 +0000 UTC m=+5453.841084206" Nov 24 08:18:29 crc kubenswrapper[4799]: I1124 08:18:29.064761 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 08:18:30 crc kubenswrapper[4799]: I1124 08:18:30.202878 4799 generic.go:334] "Generic (PLEG): container finished" podID="80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81" containerID="dce6e19b87b53b1c51b330203a18e4aa042fc14397acf2e0e92ffe3d992713a1" exitCode=0 Nov 24 08:18:30 crc kubenswrapper[4799]: I1124 08:18:30.202956 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wvq24" event={"ID":"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81","Type":"ContainerDied","Data":"dce6e19b87b53b1c51b330203a18e4aa042fc14397acf2e0e92ffe3d992713a1"} Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.632025 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.751355 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-scripts\") pod \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.751521 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-combined-ca-bundle\") pod \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.751697 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p9b9\" (UniqueName: \"kubernetes.io/projected/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-kube-api-access-4p9b9\") pod \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.751735 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-config-data\") pod \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\" (UID: \"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81\") " Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.757009 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-scripts" (OuterVolumeSpecName: "scripts") pod "80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81" (UID: "80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.757383 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-kube-api-access-4p9b9" (OuterVolumeSpecName: "kube-api-access-4p9b9") pod "80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81" (UID: "80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81"). InnerVolumeSpecName "kube-api-access-4p9b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.782208 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81" (UID: "80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.787201 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-config-data" (OuterVolumeSpecName: "config-data") pod "80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81" (UID: "80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.854565 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p9b9\" (UniqueName: \"kubernetes.io/projected/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-kube-api-access-4p9b9\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.854615 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.854628 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:31 crc kubenswrapper[4799]: I1124 08:18:31.854640 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:32 crc kubenswrapper[4799]: I1124 08:18:32.227550 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wvq24" event={"ID":"80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81","Type":"ContainerDied","Data":"e4ad8d9ed00054196a936be5a6d2f794814a15d776c2f264c1806199adb248bc"} Nov 24 08:18:32 crc kubenswrapper[4799]: I1124 08:18:32.227604 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4ad8d9ed00054196a936be5a6d2f794814a15d776c2f264c1806199adb248bc" Nov 24 08:18:32 crc kubenswrapper[4799]: I1124 08:18:32.227626 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wvq24" Nov 24 08:18:35 crc kubenswrapper[4799]: I1124 08:18:35.960338 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 24 08:18:35 crc kubenswrapper[4799]: E1124 08:18:35.961469 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81" containerName="aodh-db-sync" Nov 24 08:18:35 crc kubenswrapper[4799]: I1124 08:18:35.961488 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81" containerName="aodh-db-sync" Nov 24 08:18:35 crc kubenswrapper[4799]: I1124 08:18:35.961745 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81" containerName="aodh-db-sync" Nov 24 08:18:35 crc kubenswrapper[4799]: I1124 08:18:35.964120 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 08:18:35 crc kubenswrapper[4799]: I1124 08:18:35.968068 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-qlvql" Nov 24 08:18:35 crc kubenswrapper[4799]: I1124 08:18:35.968364 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 24 08:18:35 crc kubenswrapper[4799]: I1124 08:18:35.968591 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 24 08:18:35 crc kubenswrapper[4799]: I1124 08:18:35.977214 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.056293 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lr7j\" (UniqueName: \"kubernetes.io/projected/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-kube-api-access-2lr7j\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.056356 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-scripts\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.056634 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-config-data\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.056710 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.060275 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-pfqpq"] Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.071516 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-pfqpq"] Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.158783 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lr7j\" (UniqueName: \"kubernetes.io/projected/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-kube-api-access-2lr7j\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.158835 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-scripts\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.158922 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-config-data\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.158961 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.168661 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-scripts\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.168798 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.169683 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-config-data\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.187562 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lr7j\" (UniqueName: \"kubernetes.io/projected/fe1b6042-8017-4fd0-b755-76c8d6af4a0b-kube-api-access-2lr7j\") pod \"aodh-0\" (UID: \"fe1b6042-8017-4fd0-b755-76c8d6af4a0b\") " pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.287595 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 08:18:36 crc kubenswrapper[4799]: I1124 08:18:36.823330 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 08:18:37 crc kubenswrapper[4799]: I1124 08:18:37.036098 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-0b82-account-create-bw62n"] Nov 24 08:18:37 crc kubenswrapper[4799]: I1124 08:18:37.047356 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-0b82-account-create-bw62n"] Nov 24 08:18:37 crc kubenswrapper[4799]: I1124 08:18:37.313019 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"fe1b6042-8017-4fd0-b755-76c8d6af4a0b","Type":"ContainerStarted","Data":"876d32f93475c7c1193e5f18432ed58ed26f5ff7c14ee34d84db69f8ad66aa28"} Nov 24 08:18:37 crc kubenswrapper[4799]: I1124 08:18:37.313340 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"fe1b6042-8017-4fd0-b755-76c8d6af4a0b","Type":"ContainerStarted","Data":"476736911ba46137ac5a990fab10e1929b1cd7483a55b8d83510fb8b7e7d4d6b"} Nov 24 08:18:37 crc kubenswrapper[4799]: I1124 08:18:37.680183 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="511a70ee-d235-48a3-813b-956c39460f01" path="/var/lib/kubelet/pods/511a70ee-d235-48a3-813b-956c39460f01/volumes" Nov 24 08:18:37 crc kubenswrapper[4799]: I1124 08:18:37.680919 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35" path="/var/lib/kubelet/pods/e8d4ba69-d6d5-434d-a20b-f87c4f3e8b35/volumes" Nov 24 08:18:38 crc kubenswrapper[4799]: I1124 08:18:38.315486 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:18:38 crc kubenswrapper[4799]: I1124 08:18:38.316194 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="proxy-httpd" containerID="cri-o://37a9cbbdd7ca4e0e9d83d34193a8ea0a846f64002da5d9826ccfb0761e1a9abb" gracePeriod=30 Nov 24 08:18:38 crc kubenswrapper[4799]: I1124 08:18:38.316389 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="sg-core" containerID="cri-o://edf175a59ce367dadd20f6a930891b369beb7b84d9e933e52460e5ddf4adea46" gracePeriod=30 Nov 24 08:18:38 crc kubenswrapper[4799]: I1124 08:18:38.316534 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="ceilometer-notification-agent" containerID="cri-o://4bbdb6b5015e0bfb5b2b32149f27e147b65b647f6917662a2f142775cd5a97d7" gracePeriod=30 Nov 24 08:18:38 crc kubenswrapper[4799]: I1124 08:18:38.317982 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="ceilometer-central-agent" containerID="cri-o://be39a2e3da5f849c1b00c4a2e3dc0df0c676e2a1fdb73c19cf1342702a847e81" gracePeriod=30 Nov 24 08:18:39 crc kubenswrapper[4799]: I1124 08:18:39.341279 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"fe1b6042-8017-4fd0-b755-76c8d6af4a0b","Type":"ContainerStarted","Data":"2fb818c55304fdf6b6647c935f8bb89e28aa70d53d71ae65f5e13b41acbe93e9"} Nov 24 08:18:39 crc kubenswrapper[4799]: I1124 08:18:39.345435 4799 generic.go:334] "Generic (PLEG): container finished" podID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerID="37a9cbbdd7ca4e0e9d83d34193a8ea0a846f64002da5d9826ccfb0761e1a9abb" exitCode=0 Nov 24 08:18:39 crc kubenswrapper[4799]: I1124 08:18:39.345463 4799 generic.go:334] "Generic (PLEG): container finished" podID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerID="edf175a59ce367dadd20f6a930891b369beb7b84d9e933e52460e5ddf4adea46" exitCode=2 Nov 24 08:18:39 crc kubenswrapper[4799]: I1124 08:18:39.345472 4799 generic.go:334] "Generic (PLEG): container finished" podID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerID="be39a2e3da5f849c1b00c4a2e3dc0df0c676e2a1fdb73c19cf1342702a847e81" exitCode=0 Nov 24 08:18:39 crc kubenswrapper[4799]: I1124 08:18:39.345493 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ea732c6-c837-4ea5-ad1f-d226502ef827","Type":"ContainerDied","Data":"37a9cbbdd7ca4e0e9d83d34193a8ea0a846f64002da5d9826ccfb0761e1a9abb"} Nov 24 08:18:39 crc kubenswrapper[4799]: I1124 08:18:39.345521 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ea732c6-c837-4ea5-ad1f-d226502ef827","Type":"ContainerDied","Data":"edf175a59ce367dadd20f6a930891b369beb7b84d9e933e52460e5ddf4adea46"} Nov 24 08:18:39 crc kubenswrapper[4799]: I1124 08:18:39.345532 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ea732c6-c837-4ea5-ad1f-d226502ef827","Type":"ContainerDied","Data":"be39a2e3da5f849c1b00c4a2e3dc0df0c676e2a1fdb73c19cf1342702a847e81"} Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.622279 4799 generic.go:334] "Generic (PLEG): container finished" podID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerID="4bbdb6b5015e0bfb5b2b32149f27e147b65b647f6917662a2f142775cd5a97d7" exitCode=0 Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.623119 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ea732c6-c837-4ea5-ad1f-d226502ef827","Type":"ContainerDied","Data":"4bbdb6b5015e0bfb5b2b32149f27e147b65b647f6917662a2f142775cd5a97d7"} Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.625646 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"fe1b6042-8017-4fd0-b755-76c8d6af4a0b","Type":"ContainerStarted","Data":"701b14573181b94758406d4bdcbb4458a907b67fd4d8cc0424058280d843817c"} Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.900711 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.985099 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-scripts\") pod \"5ea732c6-c837-4ea5-ad1f-d226502ef827\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.985154 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vl2k\" (UniqueName: \"kubernetes.io/projected/5ea732c6-c837-4ea5-ad1f-d226502ef827-kube-api-access-9vl2k\") pod \"5ea732c6-c837-4ea5-ad1f-d226502ef827\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.985235 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-combined-ca-bundle\") pod \"5ea732c6-c837-4ea5-ad1f-d226502ef827\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.985303 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-run-httpd\") pod \"5ea732c6-c837-4ea5-ad1f-d226502ef827\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.985395 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-sg-core-conf-yaml\") pod \"5ea732c6-c837-4ea5-ad1f-d226502ef827\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.985466 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-log-httpd\") pod \"5ea732c6-c837-4ea5-ad1f-d226502ef827\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.985562 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-config-data\") pod \"5ea732c6-c837-4ea5-ad1f-d226502ef827\" (UID: \"5ea732c6-c837-4ea5-ad1f-d226502ef827\") " Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.987043 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5ea732c6-c837-4ea5-ad1f-d226502ef827" (UID: "5ea732c6-c837-4ea5-ad1f-d226502ef827"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.987332 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5ea732c6-c837-4ea5-ad1f-d226502ef827" (UID: "5ea732c6-c837-4ea5-ad1f-d226502ef827"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.990083 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-scripts" (OuterVolumeSpecName: "scripts") pod "5ea732c6-c837-4ea5-ad1f-d226502ef827" (UID: "5ea732c6-c837-4ea5-ad1f-d226502ef827"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:18:41 crc kubenswrapper[4799]: I1124 08:18:41.990771 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ea732c6-c837-4ea5-ad1f-d226502ef827-kube-api-access-9vl2k" (OuterVolumeSpecName: "kube-api-access-9vl2k") pod "5ea732c6-c837-4ea5-ad1f-d226502ef827" (UID: "5ea732c6-c837-4ea5-ad1f-d226502ef827"). InnerVolumeSpecName "kube-api-access-9vl2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.013525 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5ea732c6-c837-4ea5-ad1f-d226502ef827" (UID: "5ea732c6-c837-4ea5-ad1f-d226502ef827"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.060037 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ea732c6-c837-4ea5-ad1f-d226502ef827" (UID: "5ea732c6-c837-4ea5-ad1f-d226502ef827"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.086496 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-config-data" (OuterVolumeSpecName: "config-data") pod "5ea732c6-c837-4ea5-ad1f-d226502ef827" (UID: "5ea732c6-c837-4ea5-ad1f-d226502ef827"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.088226 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.088259 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.088295 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.088311 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ea732c6-c837-4ea5-ad1f-d226502ef827-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.088323 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.088336 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ea732c6-c837-4ea5-ad1f-d226502ef827-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.088373 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vl2k\" (UniqueName: \"kubernetes.io/projected/5ea732c6-c837-4ea5-ad1f-d226502ef827-kube-api-access-9vl2k\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.659801 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ea732c6-c837-4ea5-ad1f-d226502ef827","Type":"ContainerDied","Data":"540fe33108ec2671488f930ebbe4b57f9f1e1fbbebce5d180c40cb1d1cab5b56"} Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.660155 4799 scope.go:117] "RemoveContainer" containerID="37a9cbbdd7ca4e0e9d83d34193a8ea0a846f64002da5d9826ccfb0761e1a9abb" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.659856 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.666263 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"fe1b6042-8017-4fd0-b755-76c8d6af4a0b","Type":"ContainerStarted","Data":"00c3e2badedac5da890d58d5d4ba0e9eb3e5b5d0f0baf8cf5f947ac714a43745"} Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.693080 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.718016765 podStartE2EDuration="7.693055596s" podCreationTimestamp="2025-11-24 08:18:35 +0000 UTC" firstStartedPulling="2025-11-24 08:18:36.831377632 +0000 UTC m=+5462.487360106" lastFinishedPulling="2025-11-24 08:18:41.806416463 +0000 UTC m=+5467.462398937" observedRunningTime="2025-11-24 08:18:42.685746729 +0000 UTC m=+5468.341729213" watchObservedRunningTime="2025-11-24 08:18:42.693055596 +0000 UTC m=+5468.349038070" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.709930 4799 scope.go:117] "RemoveContainer" containerID="edf175a59ce367dadd20f6a930891b369beb7b84d9e933e52460e5ddf4adea46" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.720788 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.733332 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.756158 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:18:42 crc kubenswrapper[4799]: E1124 08:18:42.756824 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="ceilometer-notification-agent" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.756841 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="ceilometer-notification-agent" Nov 24 08:18:42 crc kubenswrapper[4799]: E1124 08:18:42.756877 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="sg-core" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.756901 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="sg-core" Nov 24 08:18:42 crc kubenswrapper[4799]: E1124 08:18:42.756915 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="proxy-httpd" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.756923 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="proxy-httpd" Nov 24 08:18:42 crc kubenswrapper[4799]: E1124 08:18:42.756958 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="ceilometer-central-agent" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.756966 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="ceilometer-central-agent" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.757215 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="ceilometer-central-agent" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.757241 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="ceilometer-notification-agent" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.757248 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="proxy-httpd" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.757260 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" containerName="sg-core" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.759535 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.767605 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.767798 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.777711 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.779035 4799 scope.go:117] "RemoveContainer" containerID="4bbdb6b5015e0bfb5b2b32149f27e147b65b647f6917662a2f142775cd5a97d7" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.812680 4799 scope.go:117] "RemoveContainer" containerID="be39a2e3da5f849c1b00c4a2e3dc0df0c676e2a1fdb73c19cf1342702a847e81" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.912637 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-log-httpd\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.912921 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-run-httpd\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.912961 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.912984 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.913013 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-scripts\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.913312 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-config-data\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:42 crc kubenswrapper[4799]: I1124 08:18:42.913511 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzflj\" (UniqueName: \"kubernetes.io/projected/c872c25d-f0e4-4377-bb9c-6607a635a6c0-kube-api-access-pzflj\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.015710 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-log-httpd\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.015817 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-run-httpd\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.015835 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.015853 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.015885 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-scripts\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.015929 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-config-data\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.015960 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzflj\" (UniqueName: \"kubernetes.io/projected/c872c25d-f0e4-4377-bb9c-6607a635a6c0-kube-api-access-pzflj\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.016465 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-log-httpd\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.016501 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-run-httpd\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.021582 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.022509 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-config-data\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.023464 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-scripts\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.023477 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.033692 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzflj\" (UniqueName: \"kubernetes.io/projected/c872c25d-f0e4-4377-bb9c-6607a635a6c0-kube-api-access-pzflj\") pod \"ceilometer-0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.096974 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.622213 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.648049 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ea732c6-c837-4ea5-ad1f-d226502ef827" path="/var/lib/kubelet/pods/5ea732c6-c837-4ea5-ad1f-d226502ef827/volumes" Nov 24 08:18:43 crc kubenswrapper[4799]: I1124 08:18:43.705770 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c872c25d-f0e4-4377-bb9c-6607a635a6c0","Type":"ContainerStarted","Data":"af72107efad825e3e145259f496703a12967447f0a5ec946208e87f6edd80764"} Nov 24 08:18:44 crc kubenswrapper[4799]: I1124 08:18:44.724563 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c872c25d-f0e4-4377-bb9c-6607a635a6c0","Type":"ContainerStarted","Data":"f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815"} Nov 24 08:18:45 crc kubenswrapper[4799]: I1124 08:18:45.742122 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c872c25d-f0e4-4377-bb9c-6607a635a6c0","Type":"ContainerStarted","Data":"f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51"} Nov 24 08:18:47 crc kubenswrapper[4799]: I1124 08:18:47.770817 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c872c25d-f0e4-4377-bb9c-6607a635a6c0","Type":"ContainerStarted","Data":"4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315"} Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.045467 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-78cxl"] Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.053565 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-78cxl"] Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.333430 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-j5n76"] Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.335733 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-j5n76" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.347906 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-j5n76"] Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.446386 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95e37915-feda-469a-bcb5-983988c362e5-operator-scripts\") pod \"manila-db-create-j5n76\" (UID: \"95e37915-feda-469a-bcb5-983988c362e5\") " pod="openstack/manila-db-create-j5n76" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.446520 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb6kc\" (UniqueName: \"kubernetes.io/projected/95e37915-feda-469a-bcb5-983988c362e5-kube-api-access-qb6kc\") pod \"manila-db-create-j5n76\" (UID: \"95e37915-feda-469a-bcb5-983988c362e5\") " pod="openstack/manila-db-create-j5n76" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.450118 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-55dd-account-create-x7dhk"] Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.451747 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-55dd-account-create-x7dhk" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.455732 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.467140 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-55dd-account-create-x7dhk"] Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.547991 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbkbx\" (UniqueName: \"kubernetes.io/projected/f1025848-a743-477d-9309-0b60871a2025-kube-api-access-lbkbx\") pod \"manila-55dd-account-create-x7dhk\" (UID: \"f1025848-a743-477d-9309-0b60871a2025\") " pod="openstack/manila-55dd-account-create-x7dhk" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.548245 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb6kc\" (UniqueName: \"kubernetes.io/projected/95e37915-feda-469a-bcb5-983988c362e5-kube-api-access-qb6kc\") pod \"manila-db-create-j5n76\" (UID: \"95e37915-feda-469a-bcb5-983988c362e5\") " pod="openstack/manila-db-create-j5n76" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.548498 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1025848-a743-477d-9309-0b60871a2025-operator-scripts\") pod \"manila-55dd-account-create-x7dhk\" (UID: \"f1025848-a743-477d-9309-0b60871a2025\") " pod="openstack/manila-55dd-account-create-x7dhk" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.548775 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95e37915-feda-469a-bcb5-983988c362e5-operator-scripts\") pod \"manila-db-create-j5n76\" (UID: \"95e37915-feda-469a-bcb5-983988c362e5\") " pod="openstack/manila-db-create-j5n76" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.549728 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95e37915-feda-469a-bcb5-983988c362e5-operator-scripts\") pod \"manila-db-create-j5n76\" (UID: \"95e37915-feda-469a-bcb5-983988c362e5\") " pod="openstack/manila-db-create-j5n76" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.567436 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb6kc\" (UniqueName: \"kubernetes.io/projected/95e37915-feda-469a-bcb5-983988c362e5-kube-api-access-qb6kc\") pod \"manila-db-create-j5n76\" (UID: \"95e37915-feda-469a-bcb5-983988c362e5\") " pod="openstack/manila-db-create-j5n76" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.650218 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1025848-a743-477d-9309-0b60871a2025-operator-scripts\") pod \"manila-55dd-account-create-x7dhk\" (UID: \"f1025848-a743-477d-9309-0b60871a2025\") " pod="openstack/manila-55dd-account-create-x7dhk" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.650389 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbkbx\" (UniqueName: \"kubernetes.io/projected/f1025848-a743-477d-9309-0b60871a2025-kube-api-access-lbkbx\") pod \"manila-55dd-account-create-x7dhk\" (UID: \"f1025848-a743-477d-9309-0b60871a2025\") " pod="openstack/manila-55dd-account-create-x7dhk" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.653044 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1025848-a743-477d-9309-0b60871a2025-operator-scripts\") pod \"manila-55dd-account-create-x7dhk\" (UID: \"f1025848-a743-477d-9309-0b60871a2025\") " pod="openstack/manila-55dd-account-create-x7dhk" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.671117 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbkbx\" (UniqueName: \"kubernetes.io/projected/f1025848-a743-477d-9309-0b60871a2025-kube-api-access-lbkbx\") pod \"manila-55dd-account-create-x7dhk\" (UID: \"f1025848-a743-477d-9309-0b60871a2025\") " pod="openstack/manila-55dd-account-create-x7dhk" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.696538 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-j5n76" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.780738 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-55dd-account-create-x7dhk" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.800253 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c872c25d-f0e4-4377-bb9c-6607a635a6c0","Type":"ContainerStarted","Data":"6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee"} Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.800982 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 08:18:48 crc kubenswrapper[4799]: I1124 08:18:48.864089 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.698168259 podStartE2EDuration="6.864069308s" podCreationTimestamp="2025-11-24 08:18:42 +0000 UTC" firstStartedPulling="2025-11-24 08:18:43.629498185 +0000 UTC m=+5469.285480649" lastFinishedPulling="2025-11-24 08:18:47.795399214 +0000 UTC m=+5473.451381698" observedRunningTime="2025-11-24 08:18:48.851074369 +0000 UTC m=+5474.507056853" watchObservedRunningTime="2025-11-24 08:18:48.864069308 +0000 UTC m=+5474.520051782" Nov 24 08:18:49 crc kubenswrapper[4799]: I1124 08:18:49.187124 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-j5n76"] Nov 24 08:18:49 crc kubenswrapper[4799]: W1124 08:18:49.363637 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1025848_a743_477d_9309_0b60871a2025.slice/crio-f8eced4a17e77a148b7813c3bc428be4d7cd701a4fe1ce93428835b6dd9d9ca9 WatchSource:0}: Error finding container f8eced4a17e77a148b7813c3bc428be4d7cd701a4fe1ce93428835b6dd9d9ca9: Status 404 returned error can't find the container with id f8eced4a17e77a148b7813c3bc428be4d7cd701a4fe1ce93428835b6dd9d9ca9 Nov 24 08:18:49 crc kubenswrapper[4799]: I1124 08:18:49.366197 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-55dd-account-create-x7dhk"] Nov 24 08:18:49 crc kubenswrapper[4799]: I1124 08:18:49.644515 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6698e591-c23a-477e-890e-cddb65f66fad" path="/var/lib/kubelet/pods/6698e591-c23a-477e-890e-cddb65f66fad/volumes" Nov 24 08:18:49 crc kubenswrapper[4799]: I1124 08:18:49.813552 4799 generic.go:334] "Generic (PLEG): container finished" podID="95e37915-feda-469a-bcb5-983988c362e5" containerID="5934d736be852a953e192a06ec301201db426ef57fdfcd965b85635527c406e7" exitCode=0 Nov 24 08:18:49 crc kubenswrapper[4799]: I1124 08:18:49.813617 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-j5n76" event={"ID":"95e37915-feda-469a-bcb5-983988c362e5","Type":"ContainerDied","Data":"5934d736be852a953e192a06ec301201db426ef57fdfcd965b85635527c406e7"} Nov 24 08:18:49 crc kubenswrapper[4799]: I1124 08:18:49.813652 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-j5n76" event={"ID":"95e37915-feda-469a-bcb5-983988c362e5","Type":"ContainerStarted","Data":"e20ee5de853869f8ccd557783bbf080367818ff079cc72403b47a7d9f44514e8"} Nov 24 08:18:49 crc kubenswrapper[4799]: I1124 08:18:49.816099 4799 generic.go:334] "Generic (PLEG): container finished" podID="f1025848-a743-477d-9309-0b60871a2025" containerID="bc522741670ba4c2268e6b10472e01d059aa435d62902403a626d3477676aa2b" exitCode=0 Nov 24 08:18:49 crc kubenswrapper[4799]: I1124 08:18:49.817063 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-55dd-account-create-x7dhk" event={"ID":"f1025848-a743-477d-9309-0b60871a2025","Type":"ContainerDied","Data":"bc522741670ba4c2268e6b10472e01d059aa435d62902403a626d3477676aa2b"} Nov 24 08:18:49 crc kubenswrapper[4799]: I1124 08:18:49.817129 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-55dd-account-create-x7dhk" event={"ID":"f1025848-a743-477d-9309-0b60871a2025","Type":"ContainerStarted","Data":"f8eced4a17e77a148b7813c3bc428be4d7cd701a4fe1ce93428835b6dd9d9ca9"} Nov 24 08:18:50 crc kubenswrapper[4799]: I1124 08:18:50.400704 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:18:50 crc kubenswrapper[4799]: I1124 08:18:50.401183 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:18:50 crc kubenswrapper[4799]: I1124 08:18:50.401480 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 08:18:50 crc kubenswrapper[4799]: I1124 08:18:50.402676 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7ac03e80c7a163ff61b627bd7ec43f0aeff1db92dce99a8b10b286b57ba4f7a4"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:18:50 crc kubenswrapper[4799]: I1124 08:18:50.402756 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://7ac03e80c7a163ff61b627bd7ec43f0aeff1db92dce99a8b10b286b57ba4f7a4" gracePeriod=600 Nov 24 08:18:50 crc kubenswrapper[4799]: I1124 08:18:50.832154 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="7ac03e80c7a163ff61b627bd7ec43f0aeff1db92dce99a8b10b286b57ba4f7a4" exitCode=0 Nov 24 08:18:50 crc kubenswrapper[4799]: I1124 08:18:50.832259 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"7ac03e80c7a163ff61b627bd7ec43f0aeff1db92dce99a8b10b286b57ba4f7a4"} Nov 24 08:18:50 crc kubenswrapper[4799]: I1124 08:18:50.832512 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72"} Nov 24 08:18:50 crc kubenswrapper[4799]: I1124 08:18:50.832543 4799 scope.go:117] "RemoveContainer" containerID="0d24b3ebc20d3f695368349d60fd172d4a8ac40dcc595e6654cfe6e304d6b329" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.373336 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-55dd-account-create-x7dhk" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.383282 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-j5n76" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.553341 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb6kc\" (UniqueName: \"kubernetes.io/projected/95e37915-feda-469a-bcb5-983988c362e5-kube-api-access-qb6kc\") pod \"95e37915-feda-469a-bcb5-983988c362e5\" (UID: \"95e37915-feda-469a-bcb5-983988c362e5\") " Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.554129 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1025848-a743-477d-9309-0b60871a2025-operator-scripts\") pod \"f1025848-a743-477d-9309-0b60871a2025\" (UID: \"f1025848-a743-477d-9309-0b60871a2025\") " Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.554232 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95e37915-feda-469a-bcb5-983988c362e5-operator-scripts\") pod \"95e37915-feda-469a-bcb5-983988c362e5\" (UID: \"95e37915-feda-469a-bcb5-983988c362e5\") " Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.554276 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbkbx\" (UniqueName: \"kubernetes.io/projected/f1025848-a743-477d-9309-0b60871a2025-kube-api-access-lbkbx\") pod \"f1025848-a743-477d-9309-0b60871a2025\" (UID: \"f1025848-a743-477d-9309-0b60871a2025\") " Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.554896 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1025848-a743-477d-9309-0b60871a2025-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f1025848-a743-477d-9309-0b60871a2025" (UID: "f1025848-a743-477d-9309-0b60871a2025"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.555048 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1025848-a743-477d-9309-0b60871a2025-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.555294 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e37915-feda-469a-bcb5-983988c362e5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95e37915-feda-469a-bcb5-983988c362e5" (UID: "95e37915-feda-469a-bcb5-983988c362e5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.559559 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95e37915-feda-469a-bcb5-983988c362e5-kube-api-access-qb6kc" (OuterVolumeSpecName: "kube-api-access-qb6kc") pod "95e37915-feda-469a-bcb5-983988c362e5" (UID: "95e37915-feda-469a-bcb5-983988c362e5"). InnerVolumeSpecName "kube-api-access-qb6kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.560053 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1025848-a743-477d-9309-0b60871a2025-kube-api-access-lbkbx" (OuterVolumeSpecName: "kube-api-access-lbkbx") pod "f1025848-a743-477d-9309-0b60871a2025" (UID: "f1025848-a743-477d-9309-0b60871a2025"). InnerVolumeSpecName "kube-api-access-lbkbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.657068 4799 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95e37915-feda-469a-bcb5-983988c362e5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.657095 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbkbx\" (UniqueName: \"kubernetes.io/projected/f1025848-a743-477d-9309-0b60871a2025-kube-api-access-lbkbx\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.657104 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb6kc\" (UniqueName: \"kubernetes.io/projected/95e37915-feda-469a-bcb5-983988c362e5-kube-api-access-qb6kc\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.844979 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-55dd-account-create-x7dhk" event={"ID":"f1025848-a743-477d-9309-0b60871a2025","Type":"ContainerDied","Data":"f8eced4a17e77a148b7813c3bc428be4d7cd701a4fe1ce93428835b6dd9d9ca9"} Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.845022 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8eced4a17e77a148b7813c3bc428be4d7cd701a4fe1ce93428835b6dd9d9ca9" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.845098 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-55dd-account-create-x7dhk" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.846776 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-j5n76" event={"ID":"95e37915-feda-469a-bcb5-983988c362e5","Type":"ContainerDied","Data":"e20ee5de853869f8ccd557783bbf080367818ff079cc72403b47a7d9f44514e8"} Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.846799 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e20ee5de853869f8ccd557783bbf080367818ff079cc72403b47a7d9f44514e8" Nov 24 08:18:51 crc kubenswrapper[4799]: I1124 08:18:51.846835 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-j5n76" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.802940 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-9jmx6"] Nov 24 08:18:53 crc kubenswrapper[4799]: E1124 08:18:53.803767 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1025848-a743-477d-9309-0b60871a2025" containerName="mariadb-account-create" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.803785 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1025848-a743-477d-9309-0b60871a2025" containerName="mariadb-account-create" Nov 24 08:18:53 crc kubenswrapper[4799]: E1124 08:18:53.803826 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e37915-feda-469a-bcb5-983988c362e5" containerName="mariadb-database-create" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.803832 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e37915-feda-469a-bcb5-983988c362e5" containerName="mariadb-database-create" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.804245 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1025848-a743-477d-9309-0b60871a2025" containerName="mariadb-account-create" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.804261 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e37915-feda-469a-bcb5-983988c362e5" containerName="mariadb-database-create" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.805162 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.809368 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-sgfcr" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.809442 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.832082 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-9jmx6"] Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.902945 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkc8n\" (UniqueName: \"kubernetes.io/projected/76125ed1-4d2e-4fe8-9f73-c66425bd8506-kube-api-access-lkc8n\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.903006 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-config-data\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.903037 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-job-config-data\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:53 crc kubenswrapper[4799]: I1124 08:18:53.903114 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-combined-ca-bundle\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:54 crc kubenswrapper[4799]: I1124 08:18:54.004663 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-combined-ca-bundle\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:54 crc kubenswrapper[4799]: I1124 08:18:54.004778 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkc8n\" (UniqueName: \"kubernetes.io/projected/76125ed1-4d2e-4fe8-9f73-c66425bd8506-kube-api-access-lkc8n\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:54 crc kubenswrapper[4799]: I1124 08:18:54.004834 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-config-data\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:54 crc kubenswrapper[4799]: I1124 08:18:54.004892 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-job-config-data\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:54 crc kubenswrapper[4799]: I1124 08:18:54.012125 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-combined-ca-bundle\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:54 crc kubenswrapper[4799]: I1124 08:18:54.012224 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-config-data\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:54 crc kubenswrapper[4799]: I1124 08:18:54.013187 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-job-config-data\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:54 crc kubenswrapper[4799]: I1124 08:18:54.019877 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkc8n\" (UniqueName: \"kubernetes.io/projected/76125ed1-4d2e-4fe8-9f73-c66425bd8506-kube-api-access-lkc8n\") pod \"manila-db-sync-9jmx6\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:54 crc kubenswrapper[4799]: I1124 08:18:54.204228 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-9jmx6" Nov 24 08:18:55 crc kubenswrapper[4799]: I1124 08:18:55.064423 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-9jmx6"] Nov 24 08:18:55 crc kubenswrapper[4799]: W1124 08:18:55.066022 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76125ed1_4d2e_4fe8_9f73_c66425bd8506.slice/crio-7deb91642afe108225bf6911308b7374b864ca5ca69d899fb098478c4fef64bb WatchSource:0}: Error finding container 7deb91642afe108225bf6911308b7374b864ca5ca69d899fb098478c4fef64bb: Status 404 returned error can't find the container with id 7deb91642afe108225bf6911308b7374b864ca5ca69d899fb098478c4fef64bb Nov 24 08:18:55 crc kubenswrapper[4799]: I1124 08:18:55.902390 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-9jmx6" event={"ID":"76125ed1-4d2e-4fe8-9f73-c66425bd8506","Type":"ContainerStarted","Data":"7deb91642afe108225bf6911308b7374b864ca5ca69d899fb098478c4fef64bb"} Nov 24 08:19:01 crc kubenswrapper[4799]: I1124 08:19:01.963255 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-9jmx6" event={"ID":"76125ed1-4d2e-4fe8-9f73-c66425bd8506","Type":"ContainerStarted","Data":"05dd68c42b18d98a47730e3a26ba6fbef0b517be3da5fea3622f62acb45c1bd4"} Nov 24 08:19:02 crc kubenswrapper[4799]: I1124 08:19:02.979764 4799 generic.go:334] "Generic (PLEG): container finished" podID="76125ed1-4d2e-4fe8-9f73-c66425bd8506" containerID="05dd68c42b18d98a47730e3a26ba6fbef0b517be3da5fea3622f62acb45c1bd4" exitCode=0 Nov 24 08:19:02 crc kubenswrapper[4799]: I1124 08:19:02.979907 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-9jmx6" event={"ID":"76125ed1-4d2e-4fe8-9f73-c66425bd8506","Type":"ContainerDied","Data":"05dd68c42b18d98a47730e3a26ba6fbef0b517be3da5fea3622f62acb45c1bd4"} Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.440440 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-9jmx6" Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.576551 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-combined-ca-bundle\") pod \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.576620 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkc8n\" (UniqueName: \"kubernetes.io/projected/76125ed1-4d2e-4fe8-9f73-c66425bd8506-kube-api-access-lkc8n\") pod \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.576883 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-config-data\") pod \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.577010 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-job-config-data\") pod \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\" (UID: \"76125ed1-4d2e-4fe8-9f73-c66425bd8506\") " Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.584471 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76125ed1-4d2e-4fe8-9f73-c66425bd8506-kube-api-access-lkc8n" (OuterVolumeSpecName: "kube-api-access-lkc8n") pod "76125ed1-4d2e-4fe8-9f73-c66425bd8506" (UID: "76125ed1-4d2e-4fe8-9f73-c66425bd8506"). InnerVolumeSpecName "kube-api-access-lkc8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.585456 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-config-data" (OuterVolumeSpecName: "config-data") pod "76125ed1-4d2e-4fe8-9f73-c66425bd8506" (UID: "76125ed1-4d2e-4fe8-9f73-c66425bd8506"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.589386 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "76125ed1-4d2e-4fe8-9f73-c66425bd8506" (UID: "76125ed1-4d2e-4fe8-9f73-c66425bd8506"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.622438 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76125ed1-4d2e-4fe8-9f73-c66425bd8506" (UID: "76125ed1-4d2e-4fe8-9f73-c66425bd8506"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.681282 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.681376 4799 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-job-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.681402 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76125ed1-4d2e-4fe8-9f73-c66425bd8506-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:04 crc kubenswrapper[4799]: I1124 08:19:04.681421 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkc8n\" (UniqueName: \"kubernetes.io/projected/76125ed1-4d2e-4fe8-9f73-c66425bd8506-kube-api-access-lkc8n\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.003423 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-9jmx6" event={"ID":"76125ed1-4d2e-4fe8-9f73-c66425bd8506","Type":"ContainerDied","Data":"7deb91642afe108225bf6911308b7374b864ca5ca69d899fb098478c4fef64bb"} Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.003472 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7deb91642afe108225bf6911308b7374b864ca5ca69d899fb098478c4fef64bb" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.003548 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-9jmx6" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.417455 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 08:19:05 crc kubenswrapper[4799]: E1124 08:19:05.417911 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76125ed1-4d2e-4fe8-9f73-c66425bd8506" containerName="manila-db-sync" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.417930 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="76125ed1-4d2e-4fe8-9f73-c66425bd8506" containerName="manila-db-sync" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.418464 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="76125ed1-4d2e-4fe8-9f73-c66425bd8506" containerName="manila-db-sync" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.420129 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.421670 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.422603 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-sgfcr" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.422872 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.424946 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.487688 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.489791 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.494680 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.500200 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.500271 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhrl8\" (UniqueName: \"kubernetes.io/projected/99a84edf-0397-4af1-9a2f-252678c9560e-kube-api-access-rhrl8\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.500367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-scripts\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.500470 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99a84edf-0397-4af1-9a2f-252678c9560e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.500508 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-config-data\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.500571 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.509319 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.540739 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.560018 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58c7c87865-pmsfp"] Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.561730 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.592405 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58c7c87865-pmsfp"] Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.601823 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhrl8\" (UniqueName: \"kubernetes.io/projected/99a84edf-0397-4af1-9a2f-252678c9560e-kube-api-access-rhrl8\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.601947 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zcsm\" (UniqueName: \"kubernetes.io/projected/02d972f3-0f96-40d9-94e0-949b605be376-kube-api-access-2zcsm\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.601985 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-scripts\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602018 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602057 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02d972f3-0f96-40d9-94e0-949b605be376-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602093 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99a84edf-0397-4af1-9a2f-252678c9560e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602119 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/02d972f3-0f96-40d9-94e0-949b605be376-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602157 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-config-data\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602207 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602231 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602269 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/02d972f3-0f96-40d9-94e0-949b605be376-ceph\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602296 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-config-data\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602368 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99a84edf-0397-4af1-9a2f-252678c9560e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.602390 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-scripts\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.613823 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.614223 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-scripts\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.618009 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-config-data\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.631788 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99a84edf-0397-4af1-9a2f-252678c9560e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.633102 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhrl8\" (UniqueName: \"kubernetes.io/projected/99a84edf-0397-4af1-9a2f-252678c9560e-kube-api-access-rhrl8\") pod \"manila-scheduler-0\" (UID: \"99a84edf-0397-4af1-9a2f-252678c9560e\") " pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.683380 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.685486 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.691250 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.697736 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704305 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-sb\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704385 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-scripts\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704434 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-dns-svc\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704479 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-nb\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704520 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zcsm\" (UniqueName: \"kubernetes.io/projected/02d972f3-0f96-40d9-94e0-949b605be376-kube-api-access-2zcsm\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704551 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-config\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704582 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704621 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02d972f3-0f96-40d9-94e0-949b605be376-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704658 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/02d972f3-0f96-40d9-94e0-949b605be376-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704728 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704772 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/02d972f3-0f96-40d9-94e0-949b605be376-ceph\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704799 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-config-data\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.704828 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sntpc\" (UniqueName: \"kubernetes.io/projected/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-kube-api-access-sntpc\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.709472 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.709556 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02d972f3-0f96-40d9-94e0-949b605be376-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.709656 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/02d972f3-0f96-40d9-94e0-949b605be376-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.727902 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-config-data\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.732572 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-scripts\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.740442 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.740499 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02d972f3-0f96-40d9-94e0-949b605be376-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.741036 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/02d972f3-0f96-40d9-94e0-949b605be376-ceph\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.742408 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zcsm\" (UniqueName: \"kubernetes.io/projected/02d972f3-0f96-40d9-94e0-949b605be376-kube-api-access-2zcsm\") pod \"manila-share-share1-0\" (UID: \"02d972f3-0f96-40d9-94e0-949b605be376\") " pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.806987 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-config\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.807079 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.807136 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0510e3b8-af09-4b89-8041-a5810b6fdad7-logs\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.807209 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0510e3b8-af09-4b89-8041-a5810b6fdad7-etc-machine-id\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.807241 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sntpc\" (UniqueName: \"kubernetes.io/projected/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-kube-api-access-sntpc\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.807260 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-config-data-custom\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.807290 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-config-data\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.807324 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-sb\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.807384 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4j6f\" (UniqueName: \"kubernetes.io/projected/0510e3b8-af09-4b89-8041-a5810b6fdad7-kube-api-access-v4j6f\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.807400 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-dns-svc\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.807465 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-scripts\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.807486 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-nb\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.808708 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-nb\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.809531 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-config\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.810292 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-dns-svc\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.811043 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-sb\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.826900 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.828280 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sntpc\" (UniqueName: \"kubernetes.io/projected/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-kube-api-access-sntpc\") pod \"dnsmasq-dns-58c7c87865-pmsfp\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.900749 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.911143 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0510e3b8-af09-4b89-8041-a5810b6fdad7-etc-machine-id\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.911203 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-config-data-custom\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.911228 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-config-data\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.911247 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0510e3b8-af09-4b89-8041-a5810b6fdad7-etc-machine-id\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.911300 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4j6f\" (UniqueName: \"kubernetes.io/projected/0510e3b8-af09-4b89-8041-a5810b6fdad7-kube-api-access-v4j6f\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.911332 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-scripts\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.911386 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.911430 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0510e3b8-af09-4b89-8041-a5810b6fdad7-logs\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.912225 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0510e3b8-af09-4b89-8041-a5810b6fdad7-logs\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.922757 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-scripts\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.924372 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-config-data-custom\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.925087 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-config-data\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.928413 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0510e3b8-af09-4b89-8041-a5810b6fdad7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:05 crc kubenswrapper[4799]: I1124 08:19:05.935309 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4j6f\" (UniqueName: \"kubernetes.io/projected/0510e3b8-af09-4b89-8041-a5810b6fdad7-kube-api-access-v4j6f\") pod \"manila-api-0\" (UID: \"0510e3b8-af09-4b89-8041-a5810b6fdad7\") " pod="openstack/manila-api-0" Nov 24 08:19:06 crc kubenswrapper[4799]: I1124 08:19:06.158552 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 08:19:06 crc kubenswrapper[4799]: I1124 08:19:06.368399 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 08:19:06 crc kubenswrapper[4799]: W1124 08:19:06.517972 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fff3acf_3ded_4637_96cf_e3bd8dd9eae7.slice/crio-bdba626c0b04a32966d377153a587c3b4f529366d869809a73be5b607d32f9e2 WatchSource:0}: Error finding container bdba626c0b04a32966d377153a587c3b4f529366d869809a73be5b607d32f9e2: Status 404 returned error can't find the container with id bdba626c0b04a32966d377153a587c3b4f529366d869809a73be5b607d32f9e2 Nov 24 08:19:06 crc kubenswrapper[4799]: I1124 08:19:06.518036 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58c7c87865-pmsfp"] Nov 24 08:19:06 crc kubenswrapper[4799]: I1124 08:19:06.611428 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 08:19:06 crc kubenswrapper[4799]: I1124 08:19:06.778074 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 08:19:07 crc kubenswrapper[4799]: I1124 08:19:07.025304 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0510e3b8-af09-4b89-8041-a5810b6fdad7","Type":"ContainerStarted","Data":"b928b9b5bbefdb39e0f2027fe24a0c4783d218bd0c33fd9d53e45b169a1bf6c4"} Nov 24 08:19:07 crc kubenswrapper[4799]: I1124 08:19:07.031821 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" event={"ID":"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7","Type":"ContainerStarted","Data":"c01734f24717e1a2a8ffbe43cbf95dfef75015629bc8b4d002409496ef8a3b0f"} Nov 24 08:19:07 crc kubenswrapper[4799]: I1124 08:19:07.031888 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" event={"ID":"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7","Type":"ContainerStarted","Data":"bdba626c0b04a32966d377153a587c3b4f529366d869809a73be5b607d32f9e2"} Nov 24 08:19:07 crc kubenswrapper[4799]: I1124 08:19:07.043199 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"02d972f3-0f96-40d9-94e0-949b605be376","Type":"ContainerStarted","Data":"1e470373e1f86bdc920aca9551f1c08fa819dd03d2bb47d5ac047d4d8a424ebf"} Nov 24 08:19:07 crc kubenswrapper[4799]: I1124 08:19:07.048018 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"99a84edf-0397-4af1-9a2f-252678c9560e","Type":"ContainerStarted","Data":"bf5f17195c8db3a29cbee590495a2a5ee3b029eda2401fb333d815da2168b4d0"} Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.090660 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"99a84edf-0397-4af1-9a2f-252678c9560e","Type":"ContainerStarted","Data":"d5d717fdc5be73776ffadb5f08c9aaa4ed95682e73535143d953c7ed8e23718c"} Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.091345 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"99a84edf-0397-4af1-9a2f-252678c9560e","Type":"ContainerStarted","Data":"9f40c9d9a5c120b6c14caf35b2b08fa9899ba7ccf50d494b5336fec3a82f3e61"} Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.094688 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0510e3b8-af09-4b89-8041-a5810b6fdad7","Type":"ContainerStarted","Data":"46fdd9d2f437e9238d7237a02f47ab0e394024dbc4441ce976e71ffa7b56cc03"} Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.094742 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0510e3b8-af09-4b89-8041-a5810b6fdad7","Type":"ContainerStarted","Data":"8fe95bad72bd1a844d5d3261e3fa31036f5051e61aed8ef0713b4c7347b64bc3"} Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.095804 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.098828 4799 generic.go:334] "Generic (PLEG): container finished" podID="2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" containerID="c01734f24717e1a2a8ffbe43cbf95dfef75015629bc8b4d002409496ef8a3b0f" exitCode=0 Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.098937 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" event={"ID":"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7","Type":"ContainerDied","Data":"c01734f24717e1a2a8ffbe43cbf95dfef75015629bc8b4d002409496ef8a3b0f"} Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.116470 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.844090879 podStartE2EDuration="3.116452586s" podCreationTimestamp="2025-11-24 08:19:05 +0000 UTC" firstStartedPulling="2025-11-24 08:19:06.384311206 +0000 UTC m=+5492.040293680" lastFinishedPulling="2025-11-24 08:19:06.656672913 +0000 UTC m=+5492.312655387" observedRunningTime="2025-11-24 08:19:08.11378216 +0000 UTC m=+5493.769764634" watchObservedRunningTime="2025-11-24 08:19:08.116452586 +0000 UTC m=+5493.772435060" Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.186153 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.186135175 podStartE2EDuration="3.186135175s" podCreationTimestamp="2025-11-24 08:19:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:19:08.166451696 +0000 UTC m=+5493.822434170" watchObservedRunningTime="2025-11-24 08:19:08.186135175 +0000 UTC m=+5493.842117639" Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.551728 4799 scope.go:117] "RemoveContainer" containerID="05f6af4fec51a400e21d3d540f4c963b3f3946bbec60ea13ccf30c96220851c6" Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.600140 4799 scope.go:117] "RemoveContainer" containerID="35da31ae2944ee1471c48bdc7ba0ebee08a2b4dcaeefc303180628ae20ab14fb" Nov 24 08:19:08 crc kubenswrapper[4799]: I1124 08:19:08.655879 4799 scope.go:117] "RemoveContainer" containerID="22bb21315f2a8566eadbc6d049042a8f6fc00e6b9e79af02149377c148f6dd7d" Nov 24 08:19:09 crc kubenswrapper[4799]: I1124 08:19:09.118697 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" event={"ID":"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7","Type":"ContainerStarted","Data":"ab53fdd9f9b204e912f86ed376e79c87b6376f539904e3efefd7950ff13a8658"} Nov 24 08:19:09 crc kubenswrapper[4799]: I1124 08:19:09.118917 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:09 crc kubenswrapper[4799]: I1124 08:19:09.148798 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" podStartSLOduration=4.148762228 podStartE2EDuration="4.148762228s" podCreationTimestamp="2025-11-24 08:19:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:19:09.143705964 +0000 UTC m=+5494.799688438" watchObservedRunningTime="2025-11-24 08:19:09.148762228 +0000 UTC m=+5494.804744702" Nov 24 08:19:13 crc kubenswrapper[4799]: I1124 08:19:13.110875 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 08:19:15 crc kubenswrapper[4799]: I1124 08:19:15.178348 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"02d972f3-0f96-40d9-94e0-949b605be376","Type":"ContainerStarted","Data":"1cfe11f2fef546325bf349b519083216ffef0093fed8ef77e73f79e8dee058ef"} Nov 24 08:19:15 crc kubenswrapper[4799]: I1124 08:19:15.741573 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 24 08:19:15 crc kubenswrapper[4799]: I1124 08:19:15.903139 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.005469 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b86c965c-lhnzm"] Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.005812 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" podUID="27b56ac5-d465-4341-a689-bd7c4dc5438c" containerName="dnsmasq-dns" containerID="cri-o://2c247ff0d900ac39728dae153b1a08256c5ba8196fad15da6205848517a7f697" gracePeriod=10 Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.192924 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"02d972f3-0f96-40d9-94e0-949b605be376","Type":"ContainerStarted","Data":"f2dc2779683db9d1d58cd5192153f22e4ce2280cac23268b7bfe83a0af3de5d2"} Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.196716 4799 generic.go:334] "Generic (PLEG): container finished" podID="27b56ac5-d465-4341-a689-bd7c4dc5438c" containerID="2c247ff0d900ac39728dae153b1a08256c5ba8196fad15da6205848517a7f697" exitCode=0 Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.196756 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" event={"ID":"27b56ac5-d465-4341-a689-bd7c4dc5438c","Type":"ContainerDied","Data":"2c247ff0d900ac39728dae153b1a08256c5ba8196fad15da6205848517a7f697"} Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.645442 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.672278 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.989415191 podStartE2EDuration="11.672252635s" podCreationTimestamp="2025-11-24 08:19:05 +0000 UTC" firstStartedPulling="2025-11-24 08:19:06.653637436 +0000 UTC m=+5492.309619910" lastFinishedPulling="2025-11-24 08:19:14.33647488 +0000 UTC m=+5499.992457354" observedRunningTime="2025-11-24 08:19:16.216431888 +0000 UTC m=+5501.872414382" watchObservedRunningTime="2025-11-24 08:19:16.672252635 +0000 UTC m=+5502.328235109" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.795238 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-nb\") pod \"27b56ac5-d465-4341-a689-bd7c4dc5438c\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.795352 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6jk2\" (UniqueName: \"kubernetes.io/projected/27b56ac5-d465-4341-a689-bd7c4dc5438c-kube-api-access-x6jk2\") pod \"27b56ac5-d465-4341-a689-bd7c4dc5438c\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.795709 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-dns-svc\") pod \"27b56ac5-d465-4341-a689-bd7c4dc5438c\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.795777 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-config\") pod \"27b56ac5-d465-4341-a689-bd7c4dc5438c\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.796755 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-sb\") pod \"27b56ac5-d465-4341-a689-bd7c4dc5438c\" (UID: \"27b56ac5-d465-4341-a689-bd7c4dc5438c\") " Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.819825 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27b56ac5-d465-4341-a689-bd7c4dc5438c-kube-api-access-x6jk2" (OuterVolumeSpecName: "kube-api-access-x6jk2") pod "27b56ac5-d465-4341-a689-bd7c4dc5438c" (UID: "27b56ac5-d465-4341-a689-bd7c4dc5438c"). InnerVolumeSpecName "kube-api-access-x6jk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.863939 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "27b56ac5-d465-4341-a689-bd7c4dc5438c" (UID: "27b56ac5-d465-4341-a689-bd7c4dc5438c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.868756 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-config" (OuterVolumeSpecName: "config") pod "27b56ac5-d465-4341-a689-bd7c4dc5438c" (UID: "27b56ac5-d465-4341-a689-bd7c4dc5438c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.870088 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "27b56ac5-d465-4341-a689-bd7c4dc5438c" (UID: "27b56ac5-d465-4341-a689-bd7c4dc5438c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.875388 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "27b56ac5-d465-4341-a689-bd7c4dc5438c" (UID: "27b56ac5-d465-4341-a689-bd7c4dc5438c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.899578 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6jk2\" (UniqueName: \"kubernetes.io/projected/27b56ac5-d465-4341-a689-bd7c4dc5438c-kube-api-access-x6jk2\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.899624 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.899637 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.899658 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:16 crc kubenswrapper[4799]: I1124 08:19:16.899670 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27b56ac5-d465-4341-a689-bd7c4dc5438c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:17 crc kubenswrapper[4799]: I1124 08:19:17.208670 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" Nov 24 08:19:17 crc kubenswrapper[4799]: I1124 08:19:17.210908 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b86c965c-lhnzm" event={"ID":"27b56ac5-d465-4341-a689-bd7c4dc5438c","Type":"ContainerDied","Data":"104af59ae6f184eb771eeaf90760ac5398d8db3830602a0d9c2e8d4e1e18e5cb"} Nov 24 08:19:17 crc kubenswrapper[4799]: I1124 08:19:17.210962 4799 scope.go:117] "RemoveContainer" containerID="2c247ff0d900ac39728dae153b1a08256c5ba8196fad15da6205848517a7f697" Nov 24 08:19:17 crc kubenswrapper[4799]: I1124 08:19:17.254782 4799 scope.go:117] "RemoveContainer" containerID="5ad23e59167853cdf35ee4240b3de589d0fd11b88d2a7f5e8042bdbad847ddd1" Nov 24 08:19:17 crc kubenswrapper[4799]: I1124 08:19:17.274135 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b86c965c-lhnzm"] Nov 24 08:19:17 crc kubenswrapper[4799]: I1124 08:19:17.292199 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b86c965c-lhnzm"] Nov 24 08:19:17 crc kubenswrapper[4799]: I1124 08:19:17.645331 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27b56ac5-d465-4341-a689-bd7c4dc5438c" path="/var/lib/kubelet/pods/27b56ac5-d465-4341-a689-bd7c4dc5438c/volumes" Nov 24 08:19:18 crc kubenswrapper[4799]: I1124 08:19:18.431058 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:19:18 crc kubenswrapper[4799]: I1124 08:19:18.431718 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="ceilometer-central-agent" containerID="cri-o://f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815" gracePeriod=30 Nov 24 08:19:18 crc kubenswrapper[4799]: I1124 08:19:18.431870 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="ceilometer-notification-agent" containerID="cri-o://f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51" gracePeriod=30 Nov 24 08:19:18 crc kubenswrapper[4799]: I1124 08:19:18.431883 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="sg-core" containerID="cri-o://4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315" gracePeriod=30 Nov 24 08:19:18 crc kubenswrapper[4799]: I1124 08:19:18.432089 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="proxy-httpd" containerID="cri-o://6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee" gracePeriod=30 Nov 24 08:19:19 crc kubenswrapper[4799]: I1124 08:19:19.239150 4799 generic.go:334] "Generic (PLEG): container finished" podID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerID="6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee" exitCode=0 Nov 24 08:19:19 crc kubenswrapper[4799]: I1124 08:19:19.239472 4799 generic.go:334] "Generic (PLEG): container finished" podID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerID="4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315" exitCode=2 Nov 24 08:19:19 crc kubenswrapper[4799]: I1124 08:19:19.239481 4799 generic.go:334] "Generic (PLEG): container finished" podID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerID="f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815" exitCode=0 Nov 24 08:19:19 crc kubenswrapper[4799]: I1124 08:19:19.239223 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c872c25d-f0e4-4377-bb9c-6607a635a6c0","Type":"ContainerDied","Data":"6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee"} Nov 24 08:19:19 crc kubenswrapper[4799]: I1124 08:19:19.239516 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c872c25d-f0e4-4377-bb9c-6607a635a6c0","Type":"ContainerDied","Data":"4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315"} Nov 24 08:19:19 crc kubenswrapper[4799]: I1124 08:19:19.239532 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c872c25d-f0e4-4377-bb9c-6607a635a6c0","Type":"ContainerDied","Data":"f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815"} Nov 24 08:19:21 crc kubenswrapper[4799]: I1124 08:19:21.889331 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.092463 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-combined-ca-bundle\") pod \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.092688 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-scripts\") pod \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.092737 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzflj\" (UniqueName: \"kubernetes.io/projected/c872c25d-f0e4-4377-bb9c-6607a635a6c0-kube-api-access-pzflj\") pod \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.092767 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-sg-core-conf-yaml\") pod \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.092787 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-config-data\") pod \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.092816 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-run-httpd\") pod \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.092837 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-log-httpd\") pod \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\" (UID: \"c872c25d-f0e4-4377-bb9c-6607a635a6c0\") " Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.093514 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c872c25d-f0e4-4377-bb9c-6607a635a6c0" (UID: "c872c25d-f0e4-4377-bb9c-6607a635a6c0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.093648 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c872c25d-f0e4-4377-bb9c-6607a635a6c0" (UID: "c872c25d-f0e4-4377-bb9c-6607a635a6c0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.100917 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-scripts" (OuterVolumeSpecName: "scripts") pod "c872c25d-f0e4-4377-bb9c-6607a635a6c0" (UID: "c872c25d-f0e4-4377-bb9c-6607a635a6c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.106264 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c872c25d-f0e4-4377-bb9c-6607a635a6c0-kube-api-access-pzflj" (OuterVolumeSpecName: "kube-api-access-pzflj") pod "c872c25d-f0e4-4377-bb9c-6607a635a6c0" (UID: "c872c25d-f0e4-4377-bb9c-6607a635a6c0"). InnerVolumeSpecName "kube-api-access-pzflj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.128553 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c872c25d-f0e4-4377-bb9c-6607a635a6c0" (UID: "c872c25d-f0e4-4377-bb9c-6607a635a6c0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.185698 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c872c25d-f0e4-4377-bb9c-6607a635a6c0" (UID: "c872c25d-f0e4-4377-bb9c-6607a635a6c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.195423 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.195458 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzflj\" (UniqueName: \"kubernetes.io/projected/c872c25d-f0e4-4377-bb9c-6607a635a6c0-kube-api-access-pzflj\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.195471 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.195481 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.195492 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c872c25d-f0e4-4377-bb9c-6607a635a6c0-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.195504 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.218913 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-config-data" (OuterVolumeSpecName: "config-data") pod "c872c25d-f0e4-4377-bb9c-6607a635a6c0" (UID: "c872c25d-f0e4-4377-bb9c-6607a635a6c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.297167 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c872c25d-f0e4-4377-bb9c-6607a635a6c0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.298518 4799 generic.go:334] "Generic (PLEG): container finished" podID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerID="f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51" exitCode=0 Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.298569 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c872c25d-f0e4-4377-bb9c-6607a635a6c0","Type":"ContainerDied","Data":"f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51"} Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.298607 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c872c25d-f0e4-4377-bb9c-6607a635a6c0","Type":"ContainerDied","Data":"af72107efad825e3e145259f496703a12967447f0a5ec946208e87f6edd80764"} Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.298608 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.298628 4799 scope.go:117] "RemoveContainer" containerID="6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.324971 4799 scope.go:117] "RemoveContainer" containerID="4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.343833 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.351272 4799 scope.go:117] "RemoveContainer" containerID="f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.385700 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.400784 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:19:22 crc kubenswrapper[4799]: E1124 08:19:22.401913 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27b56ac5-d465-4341-a689-bd7c4dc5438c" containerName="init" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.401929 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="27b56ac5-d465-4341-a689-bd7c4dc5438c" containerName="init" Nov 24 08:19:22 crc kubenswrapper[4799]: E1124 08:19:22.401959 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="ceilometer-notification-agent" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.401967 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="ceilometer-notification-agent" Nov 24 08:19:22 crc kubenswrapper[4799]: E1124 08:19:22.401981 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="sg-core" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.401987 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="sg-core" Nov 24 08:19:22 crc kubenswrapper[4799]: E1124 08:19:22.402010 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="proxy-httpd" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.402016 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="proxy-httpd" Nov 24 08:19:22 crc kubenswrapper[4799]: E1124 08:19:22.402035 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27b56ac5-d465-4341-a689-bd7c4dc5438c" containerName="dnsmasq-dns" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.402041 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="27b56ac5-d465-4341-a689-bd7c4dc5438c" containerName="dnsmasq-dns" Nov 24 08:19:22 crc kubenswrapper[4799]: E1124 08:19:22.402068 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="ceilometer-central-agent" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.402074 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="ceilometer-central-agent" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.402769 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="sg-core" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.402791 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="27b56ac5-d465-4341-a689-bd7c4dc5438c" containerName="dnsmasq-dns" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.402808 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="ceilometer-central-agent" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.402828 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="proxy-httpd" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.402863 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" containerName="ceilometer-notification-agent" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.404316 4799 scope.go:117] "RemoveContainer" containerID="f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.408302 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.409089 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.411297 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.412239 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.489250 4799 scope.go:117] "RemoveContainer" containerID="6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee" Nov 24 08:19:22 crc kubenswrapper[4799]: E1124 08:19:22.492214 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee\": container with ID starting with 6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee not found: ID does not exist" containerID="6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.492255 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee"} err="failed to get container status \"6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee\": rpc error: code = NotFound desc = could not find container \"6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee\": container with ID starting with 6a687fa2f648a4793a598798331327306c73f66cc22d5593473025bfbd5345ee not found: ID does not exist" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.492282 4799 scope.go:117] "RemoveContainer" containerID="4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315" Nov 24 08:19:22 crc kubenswrapper[4799]: E1124 08:19:22.493803 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315\": container with ID starting with 4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315 not found: ID does not exist" containerID="4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.493860 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315"} err="failed to get container status \"4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315\": rpc error: code = NotFound desc = could not find container \"4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315\": container with ID starting with 4df3a814c9a1c5e86efcb4d5c788f60a2ba7a22450655784616a26f35500f315 not found: ID does not exist" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.493883 4799 scope.go:117] "RemoveContainer" containerID="f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51" Nov 24 08:19:22 crc kubenswrapper[4799]: E1124 08:19:22.494406 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51\": container with ID starting with f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51 not found: ID does not exist" containerID="f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.494449 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51"} err="failed to get container status \"f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51\": rpc error: code = NotFound desc = could not find container \"f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51\": container with ID starting with f572ba94629be44b1e48cac883e00364cdbea3d2a5f486276f5f054c25313c51 not found: ID does not exist" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.494480 4799 scope.go:117] "RemoveContainer" containerID="f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815" Nov 24 08:19:22 crc kubenswrapper[4799]: E1124 08:19:22.494943 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815\": container with ID starting with f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815 not found: ID does not exist" containerID="f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.494978 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815"} err="failed to get container status \"f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815\": rpc error: code = NotFound desc = could not find container \"f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815\": container with ID starting with f1c61dc6dbf3de4cf5c86fda981022294a52189937206e0e99a4eb986e408815 not found: ID does not exist" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.604630 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-config-data\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.606222 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.606726 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06a31c2c-bbbc-4f20-94ce-9575256cf110-log-httpd\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.606888 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06a31c2c-bbbc-4f20-94ce-9575256cf110-run-httpd\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.607146 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8ltp\" (UniqueName: \"kubernetes.io/projected/06a31c2c-bbbc-4f20-94ce-9575256cf110-kube-api-access-x8ltp\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.607269 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.607360 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-scripts\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.709897 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-config-data\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.709947 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.710005 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06a31c2c-bbbc-4f20-94ce-9575256cf110-log-httpd\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.710051 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06a31c2c-bbbc-4f20-94ce-9575256cf110-run-httpd\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.710133 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8ltp\" (UniqueName: \"kubernetes.io/projected/06a31c2c-bbbc-4f20-94ce-9575256cf110-kube-api-access-x8ltp\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.710173 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.710198 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-scripts\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.711046 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06a31c2c-bbbc-4f20-94ce-9575256cf110-run-httpd\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.711146 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06a31c2c-bbbc-4f20-94ce-9575256cf110-log-httpd\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.715756 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-scripts\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.716457 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-config-data\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.716742 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.728376 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06a31c2c-bbbc-4f20-94ce-9575256cf110-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.730437 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8ltp\" (UniqueName: \"kubernetes.io/projected/06a31c2c-bbbc-4f20-94ce-9575256cf110-kube-api-access-x8ltp\") pod \"ceilometer-0\" (UID: \"06a31c2c-bbbc-4f20-94ce-9575256cf110\") " pod="openstack/ceilometer-0" Nov 24 08:19:22 crc kubenswrapper[4799]: I1124 08:19:22.771306 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:19:23 crc kubenswrapper[4799]: W1124 08:19:23.263409 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06a31c2c_bbbc_4f20_94ce_9575256cf110.slice/crio-4118bdcb5688a3f438ce15633bafee01472d1ce4e43ba2371daa1895cc3e7f7a WatchSource:0}: Error finding container 4118bdcb5688a3f438ce15633bafee01472d1ce4e43ba2371daa1895cc3e7f7a: Status 404 returned error can't find the container with id 4118bdcb5688a3f438ce15633bafee01472d1ce4e43ba2371daa1895cc3e7f7a Nov 24 08:19:23 crc kubenswrapper[4799]: I1124 08:19:23.266880 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:19:23 crc kubenswrapper[4799]: I1124 08:19:23.312680 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06a31c2c-bbbc-4f20-94ce-9575256cf110","Type":"ContainerStarted","Data":"4118bdcb5688a3f438ce15633bafee01472d1ce4e43ba2371daa1895cc3e7f7a"} Nov 24 08:19:23 crc kubenswrapper[4799]: I1124 08:19:23.643507 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c872c25d-f0e4-4377-bb9c-6607a635a6c0" path="/var/lib/kubelet/pods/c872c25d-f0e4-4377-bb9c-6607a635a6c0/volumes" Nov 24 08:19:24 crc kubenswrapper[4799]: I1124 08:19:24.327245 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06a31c2c-bbbc-4f20-94ce-9575256cf110","Type":"ContainerStarted","Data":"ac0390e65ba4a342c44e299cfefc7cd83e22331c3dc703008c3a25c6b9d2e1b7"} Nov 24 08:19:24 crc kubenswrapper[4799]: I1124 08:19:24.327580 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06a31c2c-bbbc-4f20-94ce-9575256cf110","Type":"ContainerStarted","Data":"66c0cae77cafb907e1df0a2b95546bcc68297671fa4af93a2332726a360b1a00"} Nov 24 08:19:25 crc kubenswrapper[4799]: I1124 08:19:25.339828 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06a31c2c-bbbc-4f20-94ce-9575256cf110","Type":"ContainerStarted","Data":"53c01446311964912a57b3bf30814a3455dc62453be3cac772bcb27e8a332f5c"} Nov 24 08:19:25 crc kubenswrapper[4799]: I1124 08:19:25.827785 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 24 08:19:26 crc kubenswrapper[4799]: I1124 08:19:26.356302 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06a31c2c-bbbc-4f20-94ce-9575256cf110","Type":"ContainerStarted","Data":"c42fae3addc0ce269e95a38927282dc440605bbbe40f453b22d1f81d220a628b"} Nov 24 08:19:26 crc kubenswrapper[4799]: I1124 08:19:26.356621 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 08:19:26 crc kubenswrapper[4799]: I1124 08:19:26.390050 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.856128327 podStartE2EDuration="4.390032749s" podCreationTimestamp="2025-11-24 08:19:22 +0000 UTC" firstStartedPulling="2025-11-24 08:19:23.266705344 +0000 UTC m=+5508.922687808" lastFinishedPulling="2025-11-24 08:19:25.800609756 +0000 UTC m=+5511.456592230" observedRunningTime="2025-11-24 08:19:26.38374274 +0000 UTC m=+5512.039725214" watchObservedRunningTime="2025-11-24 08:19:26.390032749 +0000 UTC m=+5512.046015223" Nov 24 08:19:27 crc kubenswrapper[4799]: I1124 08:19:27.286037 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 24 08:19:27 crc kubenswrapper[4799]: I1124 08:19:27.475352 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 24 08:19:27 crc kubenswrapper[4799]: I1124 08:19:27.584513 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Nov 24 08:19:45 crc kubenswrapper[4799]: I1124 08:19:45.052293 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-846e-account-create-x6p7t"] Nov 24 08:19:45 crc kubenswrapper[4799]: I1124 08:19:45.062102 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-ms9fw"] Nov 24 08:19:45 crc kubenswrapper[4799]: I1124 08:19:45.070255 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-ms9fw"] Nov 24 08:19:45 crc kubenswrapper[4799]: I1124 08:19:45.078436 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-846e-account-create-x6p7t"] Nov 24 08:19:45 crc kubenswrapper[4799]: I1124 08:19:45.643731 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16703067-39c9-444d-a8d3-f17344f8fe3c" path="/var/lib/kubelet/pods/16703067-39c9-444d-a8d3-f17344f8fe3c/volumes" Nov 24 08:19:45 crc kubenswrapper[4799]: I1124 08:19:45.644380 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aacd0722-6f99-4a60-a497-a76e96a5f27e" path="/var/lib/kubelet/pods/aacd0722-6f99-4a60-a497-a76e96a5f27e/volumes" Nov 24 08:19:46 crc kubenswrapper[4799]: I1124 08:19:46.073235 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-cqc6t"] Nov 24 08:19:46 crc kubenswrapper[4799]: I1124 08:19:46.083606 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-8bfc-account-create-vvn6w"] Nov 24 08:19:46 crc kubenswrapper[4799]: I1124 08:19:46.092116 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-cqc6t"] Nov 24 08:19:46 crc kubenswrapper[4799]: I1124 08:19:46.113812 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-cjhsm"] Nov 24 08:19:46 crc kubenswrapper[4799]: I1124 08:19:46.113873 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-8bfc-account-create-vvn6w"] Nov 24 08:19:46 crc kubenswrapper[4799]: I1124 08:19:46.113888 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-45f0-account-create-q4p2p"] Nov 24 08:19:46 crc kubenswrapper[4799]: I1124 08:19:46.126515 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-cjhsm"] Nov 24 08:19:46 crc kubenswrapper[4799]: I1124 08:19:46.128596 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-45f0-account-create-q4p2p"] Nov 24 08:19:47 crc kubenswrapper[4799]: I1124 08:19:47.647613 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d13f370-86ba-485c-9108-77b4149c9d3f" path="/var/lib/kubelet/pods/2d13f370-86ba-485c-9108-77b4149c9d3f/volumes" Nov 24 08:19:47 crc kubenswrapper[4799]: I1124 08:19:47.649246 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a18b4911-5cfb-4385-a5c5-8f7bf37605d8" path="/var/lib/kubelet/pods/a18b4911-5cfb-4385-a5c5-8f7bf37605d8/volumes" Nov 24 08:19:47 crc kubenswrapper[4799]: I1124 08:19:47.650576 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0c0c72c-85ae-4dca-8c60-58a9bef04b3c" path="/var/lib/kubelet/pods/c0c0c72c-85ae-4dca-8c60-58a9bef04b3c/volumes" Nov 24 08:19:47 crc kubenswrapper[4799]: I1124 08:19:47.651986 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0026133-9917-4f5c-a465-d55de6765574" path="/var/lib/kubelet/pods/f0026133-9917-4f5c-a465-d55de6765574/volumes" Nov 24 08:19:52 crc kubenswrapper[4799]: I1124 08:19:52.775744 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 08:20:05 crc kubenswrapper[4799]: I1124 08:20:05.039190 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nkdvl"] Nov 24 08:20:05 crc kubenswrapper[4799]: I1124 08:20:05.047753 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-nkdvl"] Nov 24 08:20:05 crc kubenswrapper[4799]: I1124 08:20:05.639580 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57439b00-6d98-498c-bc96-38dfcf395507" path="/var/lib/kubelet/pods/57439b00-6d98-498c-bc96-38dfcf395507/volumes" Nov 24 08:20:08 crc kubenswrapper[4799]: I1124 08:20:08.809564 4799 scope.go:117] "RemoveContainer" containerID="dbd871a6d8585a23df6a77bdf9df5a46ea1bbca5c517b00db597d57dd0bdb34a" Nov 24 08:20:08 crc kubenswrapper[4799]: I1124 08:20:08.865823 4799 scope.go:117] "RemoveContainer" containerID="56cdb01ffc7f232fdb2dd2b14aef6f670361800f20ecf6599b58f6ad3c208d63" Nov 24 08:20:08 crc kubenswrapper[4799]: I1124 08:20:08.888731 4799 scope.go:117] "RemoveContainer" containerID="598c7ea3ac2465ec38f036baebd45b07a5e38a00486d329f3551ab9f0b85dfb0" Nov 24 08:20:08 crc kubenswrapper[4799]: I1124 08:20:08.945126 4799 scope.go:117] "RemoveContainer" containerID="bb39807688abcd5f157e17009306b743ade6d35ed4c3156f413549594b846226" Nov 24 08:20:08 crc kubenswrapper[4799]: I1124 08:20:08.991066 4799 scope.go:117] "RemoveContainer" containerID="0e2d4a36f9e6eff46d6847b5020133267a5059afebf953b3fbdddd0a74909b92" Nov 24 08:20:09 crc kubenswrapper[4799]: I1124 08:20:09.039828 4799 scope.go:117] "RemoveContainer" containerID="44e5b6902a29ae87dbad1efa26ef3da911a1bef156b846bd59f80b2933fb7d5a" Nov 24 08:20:09 crc kubenswrapper[4799]: I1124 08:20:09.083726 4799 scope.go:117] "RemoveContainer" containerID="155e92ba0a2b32b54a807ec997bd5b62d4480d793f78e290425e3d0f77993f58" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.536520 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-844d6c8447-gk5jr"] Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.539032 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.541431 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.546372 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-844d6c8447-gk5jr"] Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.563633 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-nb\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.563804 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-config\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.563925 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-dns-svc\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.563965 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-sb\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.563990 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-openstack-cell1\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.564023 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnzd7\" (UniqueName: \"kubernetes.io/projected/a19fec21-dace-4f86-9d8f-6b49bcc430be-kube-api-access-hnzd7\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.665082 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-config\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.665172 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-dns-svc\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.665199 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-sb\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.665219 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-openstack-cell1\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.665241 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnzd7\" (UniqueName: \"kubernetes.io/projected/a19fec21-dace-4f86-9d8f-6b49bcc430be-kube-api-access-hnzd7\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.665311 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-nb\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.666701 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-openstack-cell1\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.667614 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-sb\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.667765 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-dns-svc\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.667765 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-config\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.667987 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-nb\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.688930 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnzd7\" (UniqueName: \"kubernetes.io/projected/a19fec21-dace-4f86-9d8f-6b49bcc430be-kube-api-access-hnzd7\") pod \"dnsmasq-dns-844d6c8447-gk5jr\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:14 crc kubenswrapper[4799]: I1124 08:20:14.859944 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:15 crc kubenswrapper[4799]: I1124 08:20:15.382475 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-844d6c8447-gk5jr"] Nov 24 08:20:15 crc kubenswrapper[4799]: I1124 08:20:15.867138 4799 generic.go:334] "Generic (PLEG): container finished" podID="a19fec21-dace-4f86-9d8f-6b49bcc430be" containerID="4eb655cf30825d2c546aae9992cad1e4f2a79bbd5fa14cd04efc789ca6f86272" exitCode=0 Nov 24 08:20:15 crc kubenswrapper[4799]: I1124 08:20:15.867201 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" event={"ID":"a19fec21-dace-4f86-9d8f-6b49bcc430be","Type":"ContainerDied","Data":"4eb655cf30825d2c546aae9992cad1e4f2a79bbd5fa14cd04efc789ca6f86272"} Nov 24 08:20:15 crc kubenswrapper[4799]: I1124 08:20:15.867440 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" event={"ID":"a19fec21-dace-4f86-9d8f-6b49bcc430be","Type":"ContainerStarted","Data":"877b1ec0a54b8a3e84d55e19ffe5f4a2a81f269a0f4f5ec1b1b1f66cdc93edd4"} Nov 24 08:20:16 crc kubenswrapper[4799]: I1124 08:20:16.879234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" event={"ID":"a19fec21-dace-4f86-9d8f-6b49bcc430be","Type":"ContainerStarted","Data":"1fd1259207e22fb320c451513f0237651bd41e2208d835dda8a2918955ec64c2"} Nov 24 08:20:16 crc kubenswrapper[4799]: I1124 08:20:16.879636 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:16 crc kubenswrapper[4799]: I1124 08:20:16.908559 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" podStartSLOduration=2.9085411519999997 podStartE2EDuration="2.908541152s" podCreationTimestamp="2025-11-24 08:20:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:20:16.901186793 +0000 UTC m=+5562.557169267" watchObservedRunningTime="2025-11-24 08:20:16.908541152 +0000 UTC m=+5562.564523616" Nov 24 08:20:20 crc kubenswrapper[4799]: I1124 08:20:20.048108 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-72q6b"] Nov 24 08:20:20 crc kubenswrapper[4799]: I1124 08:20:20.061626 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zs46w"] Nov 24 08:20:20 crc kubenswrapper[4799]: I1124 08:20:20.073592 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-72q6b"] Nov 24 08:20:20 crc kubenswrapper[4799]: I1124 08:20:20.084529 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zs46w"] Nov 24 08:20:21 crc kubenswrapper[4799]: I1124 08:20:21.641898 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97bf22b0-39f1-4302-947a-3e5ebbb078f5" path="/var/lib/kubelet/pods/97bf22b0-39f1-4302-947a-3e5ebbb078f5/volumes" Nov 24 08:20:21 crc kubenswrapper[4799]: I1124 08:20:21.642965 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1dc5c43-814a-4a95-91b2-75880697ab67" path="/var/lib/kubelet/pods/d1dc5c43-814a-4a95-91b2-75880697ab67/volumes" Nov 24 08:20:24 crc kubenswrapper[4799]: I1124 08:20:24.862001 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:24 crc kubenswrapper[4799]: I1124 08:20:24.938635 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58c7c87865-pmsfp"] Nov 24 08:20:24 crc kubenswrapper[4799]: I1124 08:20:24.938879 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" podUID="2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" containerName="dnsmasq-dns" containerID="cri-o://ab53fdd9f9b204e912f86ed376e79c87b6376f539904e3efefd7950ff13a8658" gracePeriod=10 Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.180819 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79466848d7-cb4rm"] Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.183102 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.197346 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79466848d7-cb4rm"] Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.249934 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-ovsdbserver-nb\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.249996 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-ovsdbserver-sb\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.250042 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-config\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.250068 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-dns-svc\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.250097 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-openstack-cell1\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.250343 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x5gm\" (UniqueName: \"kubernetes.io/projected/ac9ec906-a164-43a8-8d5c-274875acb978-kube-api-access-5x5gm\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.353609 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x5gm\" (UniqueName: \"kubernetes.io/projected/ac9ec906-a164-43a8-8d5c-274875acb978-kube-api-access-5x5gm\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.353783 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-ovsdbserver-nb\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.353837 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-ovsdbserver-sb\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.353899 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-config\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.353930 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-dns-svc\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.353963 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-openstack-cell1\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.355077 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-openstack-cell1\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.355393 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-ovsdbserver-sb\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.355525 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-ovsdbserver-nb\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.355773 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-config\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.355964 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac9ec906-a164-43a8-8d5c-274875acb978-dns-svc\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.374398 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x5gm\" (UniqueName: \"kubernetes.io/projected/ac9ec906-a164-43a8-8d5c-274875acb978-kube-api-access-5x5gm\") pod \"dnsmasq-dns-79466848d7-cb4rm\" (UID: \"ac9ec906-a164-43a8-8d5c-274875acb978\") " pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.551573 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.993093 4799 generic.go:334] "Generic (PLEG): container finished" podID="2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" containerID="ab53fdd9f9b204e912f86ed376e79c87b6376f539904e3efefd7950ff13a8658" exitCode=0 Nov 24 08:20:25 crc kubenswrapper[4799]: I1124 08:20:25.993270 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" event={"ID":"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7","Type":"ContainerDied","Data":"ab53fdd9f9b204e912f86ed376e79c87b6376f539904e3efefd7950ff13a8658"} Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.167920 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.278632 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-sb\") pod \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.278744 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-config\") pod \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.278799 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-dns-svc\") pod \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.279179 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sntpc\" (UniqueName: \"kubernetes.io/projected/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-kube-api-access-sntpc\") pod \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.279310 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-nb\") pod \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\" (UID: \"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7\") " Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.279828 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79466848d7-cb4rm"] Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.287641 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-kube-api-access-sntpc" (OuterVolumeSpecName: "kube-api-access-sntpc") pod "2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" (UID: "2fff3acf-3ded-4637-96cf-e3bd8dd9eae7"). InnerVolumeSpecName "kube-api-access-sntpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.366830 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" (UID: "2fff3acf-3ded-4637-96cf-e3bd8dd9eae7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.368751 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" (UID: "2fff3acf-3ded-4637-96cf-e3bd8dd9eae7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.380293 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" (UID: "2fff3acf-3ded-4637-96cf-e3bd8dd9eae7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.387394 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-config" (OuterVolumeSpecName: "config") pod "2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" (UID: "2fff3acf-3ded-4637-96cf-e3bd8dd9eae7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.388330 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sntpc\" (UniqueName: \"kubernetes.io/projected/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-kube-api-access-sntpc\") on node \"crc\" DevicePath \"\"" Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.388353 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.388362 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.388372 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:20:26 crc kubenswrapper[4799]: I1124 08:20:26.388382 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:20:27 crc kubenswrapper[4799]: I1124 08:20:27.006306 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" event={"ID":"2fff3acf-3ded-4637-96cf-e3bd8dd9eae7","Type":"ContainerDied","Data":"bdba626c0b04a32966d377153a587c3b4f529366d869809a73be5b607d32f9e2"} Nov 24 08:20:27 crc kubenswrapper[4799]: I1124 08:20:27.006784 4799 scope.go:117] "RemoveContainer" containerID="ab53fdd9f9b204e912f86ed376e79c87b6376f539904e3efefd7950ff13a8658" Nov 24 08:20:27 crc kubenswrapper[4799]: I1124 08:20:27.007047 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" Nov 24 08:20:27 crc kubenswrapper[4799]: I1124 08:20:27.013127 4799 generic.go:334] "Generic (PLEG): container finished" podID="ac9ec906-a164-43a8-8d5c-274875acb978" containerID="023137bf4b7a813b90d143388d9ca3a0ed416aa6da1303cc8806e3e2b177dee0" exitCode=0 Nov 24 08:20:27 crc kubenswrapper[4799]: I1124 08:20:27.013179 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79466848d7-cb4rm" event={"ID":"ac9ec906-a164-43a8-8d5c-274875acb978","Type":"ContainerDied","Data":"023137bf4b7a813b90d143388d9ca3a0ed416aa6da1303cc8806e3e2b177dee0"} Nov 24 08:20:27 crc kubenswrapper[4799]: I1124 08:20:27.013207 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79466848d7-cb4rm" event={"ID":"ac9ec906-a164-43a8-8d5c-274875acb978","Type":"ContainerStarted","Data":"3eb34416637f35716dd9759db5c07ab402f3dc84aca8fce0fc70732ee0e09e5e"} Nov 24 08:20:27 crc kubenswrapper[4799]: I1124 08:20:27.044921 4799 scope.go:117] "RemoveContainer" containerID="c01734f24717e1a2a8ffbe43cbf95dfef75015629bc8b4d002409496ef8a3b0f" Nov 24 08:20:27 crc kubenswrapper[4799]: I1124 08:20:27.079678 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58c7c87865-pmsfp"] Nov 24 08:20:27 crc kubenswrapper[4799]: I1124 08:20:27.098057 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58c7c87865-pmsfp"] Nov 24 08:20:27 crc kubenswrapper[4799]: I1124 08:20:27.642581 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" path="/var/lib/kubelet/pods/2fff3acf-3ded-4637-96cf-e3bd8dd9eae7/volumes" Nov 24 08:20:28 crc kubenswrapper[4799]: I1124 08:20:28.024713 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79466848d7-cb4rm" event={"ID":"ac9ec906-a164-43a8-8d5c-274875acb978","Type":"ContainerStarted","Data":"7e98ca426bc0ec51b30b0284f0dc226ebc53b835a91d70faa32cb82f811c966d"} Nov 24 08:20:28 crc kubenswrapper[4799]: I1124 08:20:28.025107 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:28 crc kubenswrapper[4799]: I1124 08:20:28.087697 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79466848d7-cb4rm" podStartSLOduration=3.087675194 podStartE2EDuration="3.087675194s" podCreationTimestamp="2025-11-24 08:20:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:20:28.080334416 +0000 UTC m=+5573.736316890" watchObservedRunningTime="2025-11-24 08:20:28.087675194 +0000 UTC m=+5573.743657668" Nov 24 08:20:30 crc kubenswrapper[4799]: I1124 08:20:30.904430 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58c7c87865-pmsfp" podUID="2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.127:5353: i/o timeout" Nov 24 08:20:34 crc kubenswrapper[4799]: I1124 08:20:34.056989 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-b8cxs"] Nov 24 08:20:34 crc kubenswrapper[4799]: I1124 08:20:34.068663 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-b8cxs"] Nov 24 08:20:35 crc kubenswrapper[4799]: I1124 08:20:35.553014 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79466848d7-cb4rm" Nov 24 08:20:35 crc kubenswrapper[4799]: I1124 08:20:35.642644 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a2e74b-c94d-448a-9d98-a0cc2539be3a" path="/var/lib/kubelet/pods/30a2e74b-c94d-448a-9d98-a0cc2539be3a/volumes" Nov 24 08:20:35 crc kubenswrapper[4799]: I1124 08:20:35.643680 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-844d6c8447-gk5jr"] Nov 24 08:20:35 crc kubenswrapper[4799]: I1124 08:20:35.643984 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" podUID="a19fec21-dace-4f86-9d8f-6b49bcc430be" containerName="dnsmasq-dns" containerID="cri-o://1fd1259207e22fb320c451513f0237651bd41e2208d835dda8a2918955ec64c2" gracePeriod=10 Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.138020 4799 generic.go:334] "Generic (PLEG): container finished" podID="a19fec21-dace-4f86-9d8f-6b49bcc430be" containerID="1fd1259207e22fb320c451513f0237651bd41e2208d835dda8a2918955ec64c2" exitCode=0 Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.138069 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" event={"ID":"a19fec21-dace-4f86-9d8f-6b49bcc430be","Type":"ContainerDied","Data":"1fd1259207e22fb320c451513f0237651bd41e2208d835dda8a2918955ec64c2"} Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.138096 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" event={"ID":"a19fec21-dace-4f86-9d8f-6b49bcc430be","Type":"ContainerDied","Data":"877b1ec0a54b8a3e84d55e19ffe5f4a2a81f269a0f4f5ec1b1b1f66cdc93edd4"} Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.138107 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="877b1ec0a54b8a3e84d55e19ffe5f4a2a81f269a0f4f5ec1b1b1f66cdc93edd4" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.152290 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.317573 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-sb\") pod \"a19fec21-dace-4f86-9d8f-6b49bcc430be\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.317748 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-openstack-cell1\") pod \"a19fec21-dace-4f86-9d8f-6b49bcc430be\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.317807 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-nb\") pod \"a19fec21-dace-4f86-9d8f-6b49bcc430be\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.317837 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnzd7\" (UniqueName: \"kubernetes.io/projected/a19fec21-dace-4f86-9d8f-6b49bcc430be-kube-api-access-hnzd7\") pod \"a19fec21-dace-4f86-9d8f-6b49bcc430be\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.317921 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-dns-svc\") pod \"a19fec21-dace-4f86-9d8f-6b49bcc430be\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.317972 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-config\") pod \"a19fec21-dace-4f86-9d8f-6b49bcc430be\" (UID: \"a19fec21-dace-4f86-9d8f-6b49bcc430be\") " Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.336656 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a19fec21-dace-4f86-9d8f-6b49bcc430be-kube-api-access-hnzd7" (OuterVolumeSpecName: "kube-api-access-hnzd7") pod "a19fec21-dace-4f86-9d8f-6b49bcc430be" (UID: "a19fec21-dace-4f86-9d8f-6b49bcc430be"). InnerVolumeSpecName "kube-api-access-hnzd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.381145 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "a19fec21-dace-4f86-9d8f-6b49bcc430be" (UID: "a19fec21-dace-4f86-9d8f-6b49bcc430be"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.383588 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-config" (OuterVolumeSpecName: "config") pod "a19fec21-dace-4f86-9d8f-6b49bcc430be" (UID: "a19fec21-dace-4f86-9d8f-6b49bcc430be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.390311 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a19fec21-dace-4f86-9d8f-6b49bcc430be" (UID: "a19fec21-dace-4f86-9d8f-6b49bcc430be"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.390664 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a19fec21-dace-4f86-9d8f-6b49bcc430be" (UID: "a19fec21-dace-4f86-9d8f-6b49bcc430be"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.391877 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a19fec21-dace-4f86-9d8f-6b49bcc430be" (UID: "a19fec21-dace-4f86-9d8f-6b49bcc430be"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.420721 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.420761 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.420779 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.420792 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-openstack-cell1\") on node \"crc\" DevicePath \"\"" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.420802 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a19fec21-dace-4f86-9d8f-6b49bcc430be-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:20:36 crc kubenswrapper[4799]: I1124 08:20:36.420819 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnzd7\" (UniqueName: \"kubernetes.io/projected/a19fec21-dace-4f86-9d8f-6b49bcc430be-kube-api-access-hnzd7\") on node \"crc\" DevicePath \"\"" Nov 24 08:20:37 crc kubenswrapper[4799]: I1124 08:20:37.149400 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844d6c8447-gk5jr" Nov 24 08:20:37 crc kubenswrapper[4799]: I1124 08:20:37.203230 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-844d6c8447-gk5jr"] Nov 24 08:20:37 crc kubenswrapper[4799]: I1124 08:20:37.216878 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-844d6c8447-gk5jr"] Nov 24 08:20:37 crc kubenswrapper[4799]: I1124 08:20:37.640065 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a19fec21-dace-4f86-9d8f-6b49bcc430be" path="/var/lib/kubelet/pods/a19fec21-dace-4f86-9d8f-6b49bcc430be/volumes" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.510881 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp"] Nov 24 08:20:46 crc kubenswrapper[4799]: E1124 08:20:46.511945 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a19fec21-dace-4f86-9d8f-6b49bcc430be" containerName="dnsmasq-dns" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.511960 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a19fec21-dace-4f86-9d8f-6b49bcc430be" containerName="dnsmasq-dns" Nov 24 08:20:46 crc kubenswrapper[4799]: E1124 08:20:46.511995 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" containerName="dnsmasq-dns" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.512004 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" containerName="dnsmasq-dns" Nov 24 08:20:46 crc kubenswrapper[4799]: E1124 08:20:46.512018 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" containerName="init" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.512027 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" containerName="init" Nov 24 08:20:46 crc kubenswrapper[4799]: E1124 08:20:46.512039 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a19fec21-dace-4f86-9d8f-6b49bcc430be" containerName="init" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.512046 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a19fec21-dace-4f86-9d8f-6b49bcc430be" containerName="init" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.512285 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fff3acf-3ded-4637-96cf-e3bd8dd9eae7" containerName="dnsmasq-dns" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.512310 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a19fec21-dace-4f86-9d8f-6b49bcc430be" containerName="dnsmasq-dns" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.513272 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.515887 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.516011 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.516125 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.516640 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.522779 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp"] Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.643191 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.643326 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.643396 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z95z\" (UniqueName: \"kubernetes.io/projected/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-kube-api-access-5z95z\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.643614 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.644006 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.746424 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.746474 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.746517 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z95z\" (UniqueName: \"kubernetes.io/projected/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-kube-api-access-5z95z\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.746576 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.746643 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.754619 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.754789 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.754859 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.755261 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.765537 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z95z\" (UniqueName: \"kubernetes.io/projected/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-kube-api-access-5z95z\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:46 crc kubenswrapper[4799]: I1124 08:20:46.853676 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:20:47 crc kubenswrapper[4799]: I1124 08:20:47.401898 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp"] Nov 24 08:20:47 crc kubenswrapper[4799]: I1124 08:20:47.407554 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:20:48 crc kubenswrapper[4799]: I1124 08:20:48.298364 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" event={"ID":"de1b21f0-9205-4688-b4d9-dbfc6045dfb1","Type":"ContainerStarted","Data":"24dcf0f7d422d64dcaca70ad37fb2250827ba75f7f4a93017771ba2a442116ac"} Nov 24 08:20:50 crc kubenswrapper[4799]: I1124 08:20:50.400589 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:20:50 crc kubenswrapper[4799]: I1124 08:20:50.401209 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:20:57 crc kubenswrapper[4799]: I1124 08:20:57.406304 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" event={"ID":"de1b21f0-9205-4688-b4d9-dbfc6045dfb1","Type":"ContainerStarted","Data":"86e10c61474cd1f8d5b1f0ba3b49c6d9b306376039d000e5a1aae82311e7c60a"} Nov 24 08:20:57 crc kubenswrapper[4799]: I1124 08:20:57.428002 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" podStartSLOduration=2.007152361 podStartE2EDuration="11.42798453s" podCreationTimestamp="2025-11-24 08:20:46 +0000 UTC" firstStartedPulling="2025-11-24 08:20:47.407331553 +0000 UTC m=+5593.063314027" lastFinishedPulling="2025-11-24 08:20:56.828163722 +0000 UTC m=+5602.484146196" observedRunningTime="2025-11-24 08:20:57.424728637 +0000 UTC m=+5603.080711121" watchObservedRunningTime="2025-11-24 08:20:57.42798453 +0000 UTC m=+5603.083967004" Nov 24 08:21:09 crc kubenswrapper[4799]: I1124 08:21:09.290427 4799 scope.go:117] "RemoveContainer" containerID="7eff771f603f594f1c891f04bffed0b19dce8367713050adf97f181b9a2ecf49" Nov 24 08:21:09 crc kubenswrapper[4799]: I1124 08:21:09.315094 4799 scope.go:117] "RemoveContainer" containerID="19f55051a854cef73d3e41a738abdda73b0fd616619d9dd47165d5ab2588ccdf" Nov 24 08:21:09 crc kubenswrapper[4799]: I1124 08:21:09.387632 4799 scope.go:117] "RemoveContainer" containerID="c03b6b7128b540198200d3ac04c9b57ab126615ec725dd3c258478c1e31a49ec" Nov 24 08:21:09 crc kubenswrapper[4799]: I1124 08:21:09.425526 4799 scope.go:117] "RemoveContainer" containerID="d009d073bd9d809d5a3bf3a99ec77b9e48ca66e7bc1fcf6cf11932881e2b264e" Nov 24 08:21:09 crc kubenswrapper[4799]: I1124 08:21:09.525501 4799 generic.go:334] "Generic (PLEG): container finished" podID="de1b21f0-9205-4688-b4d9-dbfc6045dfb1" containerID="86e10c61474cd1f8d5b1f0ba3b49c6d9b306376039d000e5a1aae82311e7c60a" exitCode=0 Nov 24 08:21:09 crc kubenswrapper[4799]: I1124 08:21:09.525639 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" event={"ID":"de1b21f0-9205-4688-b4d9-dbfc6045dfb1","Type":"ContainerDied","Data":"86e10c61474cd1f8d5b1f0ba3b49c6d9b306376039d000e5a1aae82311e7c60a"} Nov 24 08:21:09 crc kubenswrapper[4799]: I1124 08:21:09.626286 4799 scope.go:117] "RemoveContainer" containerID="7152c16ef8ee65b421528af336df1b62c92830cbb7762ae412eaa4fe8d89eb4b" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.038169 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.194515 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ssh-key\") pod \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.194659 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ceph\") pod \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.194751 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-pre-adoption-validation-combined-ca-bundle\") pod \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.194900 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-inventory\") pod \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.194934 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z95z\" (UniqueName: \"kubernetes.io/projected/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-kube-api-access-5z95z\") pod \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\" (UID: \"de1b21f0-9205-4688-b4d9-dbfc6045dfb1\") " Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.201094 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "de1b21f0-9205-4688-b4d9-dbfc6045dfb1" (UID: "de1b21f0-9205-4688-b4d9-dbfc6045dfb1"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.201295 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-kube-api-access-5z95z" (OuterVolumeSpecName: "kube-api-access-5z95z") pod "de1b21f0-9205-4688-b4d9-dbfc6045dfb1" (UID: "de1b21f0-9205-4688-b4d9-dbfc6045dfb1"). InnerVolumeSpecName "kube-api-access-5z95z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.201704 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ceph" (OuterVolumeSpecName: "ceph") pod "de1b21f0-9205-4688-b4d9-dbfc6045dfb1" (UID: "de1b21f0-9205-4688-b4d9-dbfc6045dfb1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.228180 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-inventory" (OuterVolumeSpecName: "inventory") pod "de1b21f0-9205-4688-b4d9-dbfc6045dfb1" (UID: "de1b21f0-9205-4688-b4d9-dbfc6045dfb1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.228248 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "de1b21f0-9205-4688-b4d9-dbfc6045dfb1" (UID: "de1b21f0-9205-4688-b4d9-dbfc6045dfb1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.296943 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.296977 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z95z\" (UniqueName: \"kubernetes.io/projected/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-kube-api-access-5z95z\") on node \"crc\" DevicePath \"\"" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.296991 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.296999 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.297009 4799 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1b21f0-9205-4688-b4d9-dbfc6045dfb1-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.552955 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" event={"ID":"de1b21f0-9205-4688-b4d9-dbfc6045dfb1","Type":"ContainerDied","Data":"24dcf0f7d422d64dcaca70ad37fb2250827ba75f7f4a93017771ba2a442116ac"} Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.552993 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp" Nov 24 08:21:11 crc kubenswrapper[4799]: I1124 08:21:11.552999 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24dcf0f7d422d64dcaca70ad37fb2250827ba75f7f4a93017771ba2a442116ac" Nov 24 08:21:17 crc kubenswrapper[4799]: I1124 08:21:17.041090 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-nzj9x"] Nov 24 08:21:17 crc kubenswrapper[4799]: I1124 08:21:17.055966 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f4e6-account-create-jtxfl"] Nov 24 08:21:17 crc kubenswrapper[4799]: I1124 08:21:17.066656 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f4e6-account-create-jtxfl"] Nov 24 08:21:17 crc kubenswrapper[4799]: I1124 08:21:17.075275 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-nzj9x"] Nov 24 08:21:17 crc kubenswrapper[4799]: I1124 08:21:17.648349 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2504a900-40b9-4f68-8a13-903dac0dd3d6" path="/var/lib/kubelet/pods/2504a900-40b9-4f68-8a13-903dac0dd3d6/volumes" Nov 24 08:21:17 crc kubenswrapper[4799]: I1124 08:21:17.649474 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caf0c85d-c501-4dde-af9a-c7d5aa13c619" path="/var/lib/kubelet/pods/caf0c85d-c501-4dde-af9a-c7d5aa13c619/volumes" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.228414 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z"] Nov 24 08:21:19 crc kubenswrapper[4799]: E1124 08:21:19.229160 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1b21f0-9205-4688-b4d9-dbfc6045dfb1" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.229179 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1b21f0-9205-4688-b4d9-dbfc6045dfb1" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.230002 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="de1b21f0-9205-4688-b4d9-dbfc6045dfb1" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.230795 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.247485 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.247759 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.247907 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.248521 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.258284 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z"] Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.369314 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.369634 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.369702 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.369733 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dszmf\" (UniqueName: \"kubernetes.io/projected/23397601-e699-4796-804a-bcf6b64f4a20-kube-api-access-dszmf\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.369761 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.471964 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.472006 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.472087 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.472134 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dszmf\" (UniqueName: \"kubernetes.io/projected/23397601-e699-4796-804a-bcf6b64f4a20-kube-api-access-dszmf\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.472164 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.477726 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.486421 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.486555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.486803 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.489785 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dszmf\" (UniqueName: \"kubernetes.io/projected/23397601-e699-4796-804a-bcf6b64f4a20-kube-api-access-dszmf\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:19 crc kubenswrapper[4799]: I1124 08:21:19.583128 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:21:20 crc kubenswrapper[4799]: I1124 08:21:20.101630 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z"] Nov 24 08:21:20 crc kubenswrapper[4799]: I1124 08:21:20.400775 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:21:20 crc kubenswrapper[4799]: I1124 08:21:20.400839 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:21:20 crc kubenswrapper[4799]: I1124 08:21:20.661256 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" event={"ID":"23397601-e699-4796-804a-bcf6b64f4a20","Type":"ContainerStarted","Data":"6dab5a0ae89040e6fec768ad0589646fe8373b08f8114ce123c45ff0a20ceef9"} Nov 24 08:21:21 crc kubenswrapper[4799]: I1124 08:21:21.673234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" event={"ID":"23397601-e699-4796-804a-bcf6b64f4a20","Type":"ContainerStarted","Data":"e750eb4dc8ddab1e5378123bdf61a1bae1c120359c4efd2311d3f7e1b049a80c"} Nov 24 08:21:21 crc kubenswrapper[4799]: I1124 08:21:21.699650 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" podStartSLOduration=2.297136071 podStartE2EDuration="2.699632304s" podCreationTimestamp="2025-11-24 08:21:19 +0000 UTC" firstStartedPulling="2025-11-24 08:21:20.10289939 +0000 UTC m=+5625.758881864" lastFinishedPulling="2025-11-24 08:21:20.505395623 +0000 UTC m=+5626.161378097" observedRunningTime="2025-11-24 08:21:21.693060057 +0000 UTC m=+5627.349042541" watchObservedRunningTime="2025-11-24 08:21:21.699632304 +0000 UTC m=+5627.355614778" Nov 24 08:21:50 crc kubenswrapper[4799]: I1124 08:21:50.400529 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:21:50 crc kubenswrapper[4799]: I1124 08:21:50.401050 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:21:50 crc kubenswrapper[4799]: I1124 08:21:50.401087 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 08:21:50 crc kubenswrapper[4799]: I1124 08:21:50.401793 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:21:50 crc kubenswrapper[4799]: I1124 08:21:50.401837 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" gracePeriod=600 Nov 24 08:21:50 crc kubenswrapper[4799]: E1124 08:21:50.555899 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:21:50 crc kubenswrapper[4799]: I1124 08:21:50.991309 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" exitCode=0 Nov 24 08:21:50 crc kubenswrapper[4799]: I1124 08:21:50.991419 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72"} Nov 24 08:21:50 crc kubenswrapper[4799]: I1124 08:21:50.991721 4799 scope.go:117] "RemoveContainer" containerID="7ac03e80c7a163ff61b627bd7ec43f0aeff1db92dce99a8b10b286b57ba4f7a4" Nov 24 08:21:50 crc kubenswrapper[4799]: I1124 08:21:50.992578 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:21:50 crc kubenswrapper[4799]: E1124 08:21:50.992965 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:22:01 crc kubenswrapper[4799]: I1124 08:22:01.056373 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-r7q8c"] Nov 24 08:22:01 crc kubenswrapper[4799]: I1124 08:22:01.064323 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-r7q8c"] Nov 24 08:22:01 crc kubenswrapper[4799]: I1124 08:22:01.640577 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45a30475-100a-4388-9022-0f1f18eacb82" path="/var/lib/kubelet/pods/45a30475-100a-4388-9022-0f1f18eacb82/volumes" Nov 24 08:22:02 crc kubenswrapper[4799]: I1124 08:22:02.629380 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:22:02 crc kubenswrapper[4799]: E1124 08:22:02.631254 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:22:09 crc kubenswrapper[4799]: I1124 08:22:09.801713 4799 scope.go:117] "RemoveContainer" containerID="477a30ce6f0d990499994aefcea2d8a135d3fa9ccf6de85de7c5144a5c393a5a" Nov 24 08:22:09 crc kubenswrapper[4799]: I1124 08:22:09.852623 4799 scope.go:117] "RemoveContainer" containerID="72714e01cd15036817855eabdf3bdb17d708dc9946e6d3c97c894c55df400e76" Nov 24 08:22:09 crc kubenswrapper[4799]: I1124 08:22:09.921121 4799 scope.go:117] "RemoveContainer" containerID="458c646435fa2aed0777b74c9c4698dc45a17fc9d18d985ee8a69a253f0768df" Nov 24 08:22:17 crc kubenswrapper[4799]: I1124 08:22:17.629100 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:22:17 crc kubenswrapper[4799]: E1124 08:22:17.629909 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:22:28 crc kubenswrapper[4799]: I1124 08:22:28.627944 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:22:28 crc kubenswrapper[4799]: E1124 08:22:28.628754 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:22:41 crc kubenswrapper[4799]: I1124 08:22:41.629461 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:22:41 crc kubenswrapper[4799]: E1124 08:22:41.630400 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:22:52 crc kubenswrapper[4799]: I1124 08:22:52.628132 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:22:52 crc kubenswrapper[4799]: E1124 08:22:52.628902 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:23:06 crc kubenswrapper[4799]: I1124 08:23:06.628701 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:23:06 crc kubenswrapper[4799]: E1124 08:23:06.629536 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:23:11 crc kubenswrapper[4799]: I1124 08:23:11.743535 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6v87v"] Nov 24 08:23:11 crc kubenswrapper[4799]: I1124 08:23:11.746462 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:11 crc kubenswrapper[4799]: I1124 08:23:11.779598 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6v87v"] Nov 24 08:23:11 crc kubenswrapper[4799]: I1124 08:23:11.916550 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-utilities\") pod \"certified-operators-6v87v\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:11 crc kubenswrapper[4799]: I1124 08:23:11.917156 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-catalog-content\") pod \"certified-operators-6v87v\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:11 crc kubenswrapper[4799]: I1124 08:23:11.917263 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppz9n\" (UniqueName: \"kubernetes.io/projected/69e1f89b-25d2-4df3-ac1a-3488de657758-kube-api-access-ppz9n\") pod \"certified-operators-6v87v\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:11 crc kubenswrapper[4799]: I1124 08:23:11.941174 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b4std"] Nov 24 08:23:11 crc kubenswrapper[4799]: I1124 08:23:11.943813 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:11 crc kubenswrapper[4799]: I1124 08:23:11.965444 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b4std"] Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.018683 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-utilities\") pod \"certified-operators-6v87v\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.018798 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-catalog-content\") pod \"certified-operators-6v87v\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.018907 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppz9n\" (UniqueName: \"kubernetes.io/projected/69e1f89b-25d2-4df3-ac1a-3488de657758-kube-api-access-ppz9n\") pod \"certified-operators-6v87v\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.019324 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-utilities\") pod \"certified-operators-6v87v\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.020683 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-catalog-content\") pod \"certified-operators-6v87v\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.039302 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppz9n\" (UniqueName: \"kubernetes.io/projected/69e1f89b-25d2-4df3-ac1a-3488de657758-kube-api-access-ppz9n\") pod \"certified-operators-6v87v\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.084326 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.121289 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-catalog-content\") pod \"redhat-operators-b4std\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.121359 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-utilities\") pod \"redhat-operators-b4std\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.121419 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvtrc\" (UniqueName: \"kubernetes.io/projected/0e85592c-6716-40e3-a772-d264d37915e5-kube-api-access-mvtrc\") pod \"redhat-operators-b4std\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.239053 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-catalog-content\") pod \"redhat-operators-b4std\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.239416 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-utilities\") pod \"redhat-operators-b4std\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.239475 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvtrc\" (UniqueName: \"kubernetes.io/projected/0e85592c-6716-40e3-a772-d264d37915e5-kube-api-access-mvtrc\") pod \"redhat-operators-b4std\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.240204 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-catalog-content\") pod \"redhat-operators-b4std\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.240457 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-utilities\") pod \"redhat-operators-b4std\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.271952 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvtrc\" (UniqueName: \"kubernetes.io/projected/0e85592c-6716-40e3-a772-d264d37915e5-kube-api-access-mvtrc\") pod \"redhat-operators-b4std\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.563499 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.625924 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6v87v"] Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.921303 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6v87v" event={"ID":"69e1f89b-25d2-4df3-ac1a-3488de657758","Type":"ContainerStarted","Data":"5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7"} Nov 24 08:23:12 crc kubenswrapper[4799]: I1124 08:23:12.921607 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6v87v" event={"ID":"69e1f89b-25d2-4df3-ac1a-3488de657758","Type":"ContainerStarted","Data":"29b134e60721448ce0cf00f38991968e001c284f77d394d48b38d5b5578196ab"} Nov 24 08:23:13 crc kubenswrapper[4799]: I1124 08:23:13.112975 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b4std"] Nov 24 08:23:13 crc kubenswrapper[4799]: I1124 08:23:13.934075 4799 generic.go:334] "Generic (PLEG): container finished" podID="0e85592c-6716-40e3-a772-d264d37915e5" containerID="f05ec037dc1858485c554eed9e1d28f01ea7ca3634a83e65b00bac94c5b379ba" exitCode=0 Nov 24 08:23:13 crc kubenswrapper[4799]: I1124 08:23:13.934208 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4std" event={"ID":"0e85592c-6716-40e3-a772-d264d37915e5","Type":"ContainerDied","Data":"f05ec037dc1858485c554eed9e1d28f01ea7ca3634a83e65b00bac94c5b379ba"} Nov 24 08:23:13 crc kubenswrapper[4799]: I1124 08:23:13.934813 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4std" event={"ID":"0e85592c-6716-40e3-a772-d264d37915e5","Type":"ContainerStarted","Data":"a278e0fbdd2a3a64d566099c088df4088c2ce8cc44cc051c5b04cdf3292eb5f5"} Nov 24 08:23:13 crc kubenswrapper[4799]: I1124 08:23:13.942697 4799 generic.go:334] "Generic (PLEG): container finished" podID="69e1f89b-25d2-4df3-ac1a-3488de657758" containerID="5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7" exitCode=0 Nov 24 08:23:13 crc kubenswrapper[4799]: I1124 08:23:13.942737 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6v87v" event={"ID":"69e1f89b-25d2-4df3-ac1a-3488de657758","Type":"ContainerDied","Data":"5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7"} Nov 24 08:23:13 crc kubenswrapper[4799]: I1124 08:23:13.942764 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6v87v" event={"ID":"69e1f89b-25d2-4df3-ac1a-3488de657758","Type":"ContainerStarted","Data":"e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f"} Nov 24 08:23:14 crc kubenswrapper[4799]: I1124 08:23:14.956690 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4std" event={"ID":"0e85592c-6716-40e3-a772-d264d37915e5","Type":"ContainerStarted","Data":"9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4"} Nov 24 08:23:15 crc kubenswrapper[4799]: I1124 08:23:15.968178 4799 generic.go:334] "Generic (PLEG): container finished" podID="69e1f89b-25d2-4df3-ac1a-3488de657758" containerID="e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f" exitCode=0 Nov 24 08:23:15 crc kubenswrapper[4799]: I1124 08:23:15.968283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6v87v" event={"ID":"69e1f89b-25d2-4df3-ac1a-3488de657758","Type":"ContainerDied","Data":"e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f"} Nov 24 08:23:16 crc kubenswrapper[4799]: I1124 08:23:16.983870 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6v87v" event={"ID":"69e1f89b-25d2-4df3-ac1a-3488de657758","Type":"ContainerStarted","Data":"69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833"} Nov 24 08:23:17 crc kubenswrapper[4799]: I1124 08:23:17.008584 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6v87v" podStartSLOduration=2.402525086 podStartE2EDuration="6.008554662s" podCreationTimestamp="2025-11-24 08:23:11 +0000 UTC" firstStartedPulling="2025-11-24 08:23:12.929030487 +0000 UTC m=+5738.585012961" lastFinishedPulling="2025-11-24 08:23:16.535060063 +0000 UTC m=+5742.191042537" observedRunningTime="2025-11-24 08:23:17.004997361 +0000 UTC m=+5742.660979835" watchObservedRunningTime="2025-11-24 08:23:17.008554662 +0000 UTC m=+5742.664537136" Nov 24 08:23:17 crc kubenswrapper[4799]: I1124 08:23:17.628513 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:23:17 crc kubenswrapper[4799]: E1124 08:23:17.628840 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:23:21 crc kubenswrapper[4799]: I1124 08:23:21.022641 4799 generic.go:334] "Generic (PLEG): container finished" podID="0e85592c-6716-40e3-a772-d264d37915e5" containerID="9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4" exitCode=0 Nov 24 08:23:21 crc kubenswrapper[4799]: I1124 08:23:21.022726 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4std" event={"ID":"0e85592c-6716-40e3-a772-d264d37915e5","Type":"ContainerDied","Data":"9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4"} Nov 24 08:23:22 crc kubenswrapper[4799]: I1124 08:23:22.036484 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4std" event={"ID":"0e85592c-6716-40e3-a772-d264d37915e5","Type":"ContainerStarted","Data":"d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d"} Nov 24 08:23:22 crc kubenswrapper[4799]: I1124 08:23:22.057404 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b4std" podStartSLOduration=3.51690376 podStartE2EDuration="11.057381939s" podCreationTimestamp="2025-11-24 08:23:11 +0000 UTC" firstStartedPulling="2025-11-24 08:23:13.937982466 +0000 UTC m=+5739.593964940" lastFinishedPulling="2025-11-24 08:23:21.478460645 +0000 UTC m=+5747.134443119" observedRunningTime="2025-11-24 08:23:22.055629009 +0000 UTC m=+5747.711611493" watchObservedRunningTime="2025-11-24 08:23:22.057381939 +0000 UTC m=+5747.713364413" Nov 24 08:23:22 crc kubenswrapper[4799]: I1124 08:23:22.084885 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:22 crc kubenswrapper[4799]: I1124 08:23:22.085097 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:22 crc kubenswrapper[4799]: I1124 08:23:22.149317 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:22 crc kubenswrapper[4799]: I1124 08:23:22.563986 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:22 crc kubenswrapper[4799]: I1124 08:23:22.564047 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:23 crc kubenswrapper[4799]: I1124 08:23:23.099692 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:23 crc kubenswrapper[4799]: I1124 08:23:23.297522 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6v87v"] Nov 24 08:23:23 crc kubenswrapper[4799]: I1124 08:23:23.626618 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b4std" podUID="0e85592c-6716-40e3-a772-d264d37915e5" containerName="registry-server" probeResult="failure" output=< Nov 24 08:23:23 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 08:23:23 crc kubenswrapper[4799]: > Nov 24 08:23:25 crc kubenswrapper[4799]: I1124 08:23:25.065343 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6v87v" podUID="69e1f89b-25d2-4df3-ac1a-3488de657758" containerName="registry-server" containerID="cri-o://69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833" gracePeriod=2 Nov 24 08:23:25 crc kubenswrapper[4799]: I1124 08:23:25.551765 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:25 crc kubenswrapper[4799]: I1124 08:23:25.579830 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppz9n\" (UniqueName: \"kubernetes.io/projected/69e1f89b-25d2-4df3-ac1a-3488de657758-kube-api-access-ppz9n\") pod \"69e1f89b-25d2-4df3-ac1a-3488de657758\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " Nov 24 08:23:25 crc kubenswrapper[4799]: I1124 08:23:25.580054 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-catalog-content\") pod \"69e1f89b-25d2-4df3-ac1a-3488de657758\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " Nov 24 08:23:25 crc kubenswrapper[4799]: I1124 08:23:25.580114 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-utilities\") pod \"69e1f89b-25d2-4df3-ac1a-3488de657758\" (UID: \"69e1f89b-25d2-4df3-ac1a-3488de657758\") " Nov 24 08:23:25 crc kubenswrapper[4799]: I1124 08:23:25.580683 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-utilities" (OuterVolumeSpecName: "utilities") pod "69e1f89b-25d2-4df3-ac1a-3488de657758" (UID: "69e1f89b-25d2-4df3-ac1a-3488de657758"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:23:25 crc kubenswrapper[4799]: I1124 08:23:25.589131 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69e1f89b-25d2-4df3-ac1a-3488de657758-kube-api-access-ppz9n" (OuterVolumeSpecName: "kube-api-access-ppz9n") pod "69e1f89b-25d2-4df3-ac1a-3488de657758" (UID: "69e1f89b-25d2-4df3-ac1a-3488de657758"). InnerVolumeSpecName "kube-api-access-ppz9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:23:25 crc kubenswrapper[4799]: I1124 08:23:25.624973 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69e1f89b-25d2-4df3-ac1a-3488de657758" (UID: "69e1f89b-25d2-4df3-ac1a-3488de657758"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:23:25 crc kubenswrapper[4799]: I1124 08:23:25.686378 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppz9n\" (UniqueName: \"kubernetes.io/projected/69e1f89b-25d2-4df3-ac1a-3488de657758-kube-api-access-ppz9n\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:25 crc kubenswrapper[4799]: I1124 08:23:25.686413 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:25 crc kubenswrapper[4799]: I1124 08:23:25.686424 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69e1f89b-25d2-4df3-ac1a-3488de657758-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.124381 4799 generic.go:334] "Generic (PLEG): container finished" podID="69e1f89b-25d2-4df3-ac1a-3488de657758" containerID="69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833" exitCode=0 Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.124463 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6v87v" event={"ID":"69e1f89b-25d2-4df3-ac1a-3488de657758","Type":"ContainerDied","Data":"69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833"} Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.124522 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6v87v" Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.124819 4799 scope.go:117] "RemoveContainer" containerID="69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833" Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.124803 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6v87v" event={"ID":"69e1f89b-25d2-4df3-ac1a-3488de657758","Type":"ContainerDied","Data":"29b134e60721448ce0cf00f38991968e001c284f77d394d48b38d5b5578196ab"} Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.154740 4799 scope.go:117] "RemoveContainer" containerID="e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f" Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.158283 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6v87v"] Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.170080 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6v87v"] Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.178649 4799 scope.go:117] "RemoveContainer" containerID="5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7" Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.238757 4799 scope.go:117] "RemoveContainer" containerID="69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833" Nov 24 08:23:26 crc kubenswrapper[4799]: E1124 08:23:26.239319 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833\": container with ID starting with 69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833 not found: ID does not exist" containerID="69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833" Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.239368 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833"} err="failed to get container status \"69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833\": rpc error: code = NotFound desc = could not find container \"69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833\": container with ID starting with 69f074135a0967484442c49b1b5df32615646b5649bb1c49bc4f310ee31d5833 not found: ID does not exist" Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.239398 4799 scope.go:117] "RemoveContainer" containerID="e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f" Nov 24 08:23:26 crc kubenswrapper[4799]: E1124 08:23:26.239717 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f\": container with ID starting with e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f not found: ID does not exist" containerID="e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f" Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.239756 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f"} err="failed to get container status \"e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f\": rpc error: code = NotFound desc = could not find container \"e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f\": container with ID starting with e603456e1ece1be21ded21ad0d54fae43f574981e79bdefb6634cbb5ab3c595f not found: ID does not exist" Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.239783 4799 scope.go:117] "RemoveContainer" containerID="5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7" Nov 24 08:23:26 crc kubenswrapper[4799]: E1124 08:23:26.240160 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7\": container with ID starting with 5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7 not found: ID does not exist" containerID="5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7" Nov 24 08:23:26 crc kubenswrapper[4799]: I1124 08:23:26.240190 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7"} err="failed to get container status \"5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7\": rpc error: code = NotFound desc = could not find container \"5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7\": container with ID starting with 5dea067f83eb7842ef6eecacd953297bfb65a7d251cbe906d4554ec24aa769d7 not found: ID does not exist" Nov 24 08:23:27 crc kubenswrapper[4799]: I1124 08:23:27.646582 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69e1f89b-25d2-4df3-ac1a-3488de657758" path="/var/lib/kubelet/pods/69e1f89b-25d2-4df3-ac1a-3488de657758/volumes" Nov 24 08:23:29 crc kubenswrapper[4799]: I1124 08:23:29.628415 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:23:29 crc kubenswrapper[4799]: E1124 08:23:29.629226 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:23:33 crc kubenswrapper[4799]: I1124 08:23:33.636170 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b4std" podUID="0e85592c-6716-40e3-a772-d264d37915e5" containerName="registry-server" probeResult="failure" output=< Nov 24 08:23:33 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 08:23:33 crc kubenswrapper[4799]: > Nov 24 08:23:42 crc kubenswrapper[4799]: I1124 08:23:42.624955 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:42 crc kubenswrapper[4799]: I1124 08:23:42.699010 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:42 crc kubenswrapper[4799]: I1124 08:23:42.949448 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b4std"] Nov 24 08:23:44 crc kubenswrapper[4799]: I1124 08:23:44.511344 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b4std" podUID="0e85592c-6716-40e3-a772-d264d37915e5" containerName="registry-server" containerID="cri-o://d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d" gracePeriod=2 Nov 24 08:23:44 crc kubenswrapper[4799]: I1124 08:23:44.629029 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:23:44 crc kubenswrapper[4799]: E1124 08:23:44.629263 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.010298 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.063490 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvtrc\" (UniqueName: \"kubernetes.io/projected/0e85592c-6716-40e3-a772-d264d37915e5-kube-api-access-mvtrc\") pod \"0e85592c-6716-40e3-a772-d264d37915e5\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.063571 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-utilities\") pod \"0e85592c-6716-40e3-a772-d264d37915e5\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.064917 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-utilities" (OuterVolumeSpecName: "utilities") pod "0e85592c-6716-40e3-a772-d264d37915e5" (UID: "0e85592c-6716-40e3-a772-d264d37915e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.071511 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e85592c-6716-40e3-a772-d264d37915e5-kube-api-access-mvtrc" (OuterVolumeSpecName: "kube-api-access-mvtrc") pod "0e85592c-6716-40e3-a772-d264d37915e5" (UID: "0e85592c-6716-40e3-a772-d264d37915e5"). InnerVolumeSpecName "kube-api-access-mvtrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.165923 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-catalog-content\") pod \"0e85592c-6716-40e3-a772-d264d37915e5\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.166642 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.166666 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvtrc\" (UniqueName: \"kubernetes.io/projected/0e85592c-6716-40e3-a772-d264d37915e5-kube-api-access-mvtrc\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.267137 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e85592c-6716-40e3-a772-d264d37915e5" (UID: "0e85592c-6716-40e3-a772-d264d37915e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.267836 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-catalog-content\") pod \"0e85592c-6716-40e3-a772-d264d37915e5\" (UID: \"0e85592c-6716-40e3-a772-d264d37915e5\") " Nov 24 08:23:45 crc kubenswrapper[4799]: W1124 08:23:45.267988 4799 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/0e85592c-6716-40e3-a772-d264d37915e5/volumes/kubernetes.io~empty-dir/catalog-content Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.268022 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e85592c-6716-40e3-a772-d264d37915e5" (UID: "0e85592c-6716-40e3-a772-d264d37915e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.268699 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e85592c-6716-40e3-a772-d264d37915e5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.526318 4799 generic.go:334] "Generic (PLEG): container finished" podID="0e85592c-6716-40e3-a772-d264d37915e5" containerID="d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d" exitCode=0 Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.526366 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4std" event={"ID":"0e85592c-6716-40e3-a772-d264d37915e5","Type":"ContainerDied","Data":"d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d"} Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.526405 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4std" event={"ID":"0e85592c-6716-40e3-a772-d264d37915e5","Type":"ContainerDied","Data":"a278e0fbdd2a3a64d566099c088df4088c2ce8cc44cc051c5b04cdf3292eb5f5"} Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.526427 4799 scope.go:117] "RemoveContainer" containerID="d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.526421 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b4std" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.566058 4799 scope.go:117] "RemoveContainer" containerID="9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.604014 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b4std"] Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.610233 4799 scope.go:117] "RemoveContainer" containerID="f05ec037dc1858485c554eed9e1d28f01ea7ca3634a83e65b00bac94c5b379ba" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.659651 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b4std"] Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.660922 4799 scope.go:117] "RemoveContainer" containerID="d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d" Nov 24 08:23:45 crc kubenswrapper[4799]: E1124 08:23:45.661536 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d\": container with ID starting with d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d not found: ID does not exist" containerID="d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.661590 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d"} err="failed to get container status \"d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d\": rpc error: code = NotFound desc = could not find container \"d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d\": container with ID starting with d50b2dab40f19c35fb85f60ff8ee9a24596153bf36841c563df17192674f397d not found: ID does not exist" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.661620 4799 scope.go:117] "RemoveContainer" containerID="9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4" Nov 24 08:23:45 crc kubenswrapper[4799]: E1124 08:23:45.662031 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4\": container with ID starting with 9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4 not found: ID does not exist" containerID="9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.662176 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4"} err="failed to get container status \"9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4\": rpc error: code = NotFound desc = could not find container \"9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4\": container with ID starting with 9dc6d0dddf6d79b2e2ae2d7b7280946377b294e0f1d61ebb7a93e9b60e7050e4 not found: ID does not exist" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.662305 4799 scope.go:117] "RemoveContainer" containerID="f05ec037dc1858485c554eed9e1d28f01ea7ca3634a83e65b00bac94c5b379ba" Nov 24 08:23:45 crc kubenswrapper[4799]: E1124 08:23:45.662747 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f05ec037dc1858485c554eed9e1d28f01ea7ca3634a83e65b00bac94c5b379ba\": container with ID starting with f05ec037dc1858485c554eed9e1d28f01ea7ca3634a83e65b00bac94c5b379ba not found: ID does not exist" containerID="f05ec037dc1858485c554eed9e1d28f01ea7ca3634a83e65b00bac94c5b379ba" Nov 24 08:23:45 crc kubenswrapper[4799]: I1124 08:23:45.662780 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f05ec037dc1858485c554eed9e1d28f01ea7ca3634a83e65b00bac94c5b379ba"} err="failed to get container status \"f05ec037dc1858485c554eed9e1d28f01ea7ca3634a83e65b00bac94c5b379ba\": rpc error: code = NotFound desc = could not find container \"f05ec037dc1858485c554eed9e1d28f01ea7ca3634a83e65b00bac94c5b379ba\": container with ID starting with f05ec037dc1858485c554eed9e1d28f01ea7ca3634a83e65b00bac94c5b379ba not found: ID does not exist" Nov 24 08:23:47 crc kubenswrapper[4799]: I1124 08:23:47.648105 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e85592c-6716-40e3-a772-d264d37915e5" path="/var/lib/kubelet/pods/0e85592c-6716-40e3-a772-d264d37915e5/volumes" Nov 24 08:23:59 crc kubenswrapper[4799]: I1124 08:23:59.629136 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:23:59 crc kubenswrapper[4799]: E1124 08:23:59.629995 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:24:10 crc kubenswrapper[4799]: I1124 08:24:10.628841 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:24:10 crc kubenswrapper[4799]: E1124 08:24:10.629617 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:24:21 crc kubenswrapper[4799]: I1124 08:24:21.629095 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:24:21 crc kubenswrapper[4799]: E1124 08:24:21.629939 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:24:34 crc kubenswrapper[4799]: I1124 08:24:34.628948 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:24:34 crc kubenswrapper[4799]: E1124 08:24:34.629780 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:24:45 crc kubenswrapper[4799]: I1124 08:24:45.635603 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:24:45 crc kubenswrapper[4799]: E1124 08:24:45.637240 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.411388 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5qf95"] Nov 24 08:24:51 crc kubenswrapper[4799]: E1124 08:24:51.418371 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e1f89b-25d2-4df3-ac1a-3488de657758" containerName="registry-server" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.418558 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e1f89b-25d2-4df3-ac1a-3488de657758" containerName="registry-server" Nov 24 08:24:51 crc kubenswrapper[4799]: E1124 08:24:51.418674 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e85592c-6716-40e3-a772-d264d37915e5" containerName="extract-content" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.418758 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e85592c-6716-40e3-a772-d264d37915e5" containerName="extract-content" Nov 24 08:24:51 crc kubenswrapper[4799]: E1124 08:24:51.418869 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e1f89b-25d2-4df3-ac1a-3488de657758" containerName="extract-utilities" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.418964 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e1f89b-25d2-4df3-ac1a-3488de657758" containerName="extract-utilities" Nov 24 08:24:51 crc kubenswrapper[4799]: E1124 08:24:51.419053 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e1f89b-25d2-4df3-ac1a-3488de657758" containerName="extract-content" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.419134 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e1f89b-25d2-4df3-ac1a-3488de657758" containerName="extract-content" Nov 24 08:24:51 crc kubenswrapper[4799]: E1124 08:24:51.419216 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e85592c-6716-40e3-a772-d264d37915e5" containerName="extract-utilities" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.419291 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e85592c-6716-40e3-a772-d264d37915e5" containerName="extract-utilities" Nov 24 08:24:51 crc kubenswrapper[4799]: E1124 08:24:51.419375 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e85592c-6716-40e3-a772-d264d37915e5" containerName="registry-server" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.419454 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e85592c-6716-40e3-a772-d264d37915e5" containerName="registry-server" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.419797 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e85592c-6716-40e3-a772-d264d37915e5" containerName="registry-server" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.419917 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="69e1f89b-25d2-4df3-ac1a-3488de657758" containerName="registry-server" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.421769 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.432040 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5qf95"] Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.564077 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-utilities\") pod \"community-operators-5qf95\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.564128 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjxvz\" (UniqueName: \"kubernetes.io/projected/749e575d-e292-4ae6-9797-a0af7a566147-kube-api-access-mjxvz\") pod \"community-operators-5qf95\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.564253 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-catalog-content\") pod \"community-operators-5qf95\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.666144 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-catalog-content\") pod \"community-operators-5qf95\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.666295 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-utilities\") pod \"community-operators-5qf95\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.666319 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjxvz\" (UniqueName: \"kubernetes.io/projected/749e575d-e292-4ae6-9797-a0af7a566147-kube-api-access-mjxvz\") pod \"community-operators-5qf95\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.666730 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-catalog-content\") pod \"community-operators-5qf95\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.666811 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-utilities\") pod \"community-operators-5qf95\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.691402 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjxvz\" (UniqueName: \"kubernetes.io/projected/749e575d-e292-4ae6-9797-a0af7a566147-kube-api-access-mjxvz\") pod \"community-operators-5qf95\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:24:51 crc kubenswrapper[4799]: I1124 08:24:51.750766 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:24:52 crc kubenswrapper[4799]: I1124 08:24:52.364100 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5qf95"] Nov 24 08:24:53 crc kubenswrapper[4799]: I1124 08:24:53.271996 4799 generic.go:334] "Generic (PLEG): container finished" podID="749e575d-e292-4ae6-9797-a0af7a566147" containerID="5c2907b0d63822cbeb2a3fab4d00ab6e475c4ec296a1b2b4ea35f40194279ee8" exitCode=0 Nov 24 08:24:53 crc kubenswrapper[4799]: I1124 08:24:53.272073 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qf95" event={"ID":"749e575d-e292-4ae6-9797-a0af7a566147","Type":"ContainerDied","Data":"5c2907b0d63822cbeb2a3fab4d00ab6e475c4ec296a1b2b4ea35f40194279ee8"} Nov 24 08:24:53 crc kubenswrapper[4799]: I1124 08:24:53.272370 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qf95" event={"ID":"749e575d-e292-4ae6-9797-a0af7a566147","Type":"ContainerStarted","Data":"56ebe70791a4bba1eb3b1267a3591e2a406ce31ad71763d7fdbc6a14ed245b84"} Nov 24 08:24:54 crc kubenswrapper[4799]: I1124 08:24:54.287898 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qf95" event={"ID":"749e575d-e292-4ae6-9797-a0af7a566147","Type":"ContainerStarted","Data":"450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d"} Nov 24 08:24:56 crc kubenswrapper[4799]: I1124 08:24:56.307869 4799 generic.go:334] "Generic (PLEG): container finished" podID="749e575d-e292-4ae6-9797-a0af7a566147" containerID="450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d" exitCode=0 Nov 24 08:24:56 crc kubenswrapper[4799]: I1124 08:24:56.307905 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qf95" event={"ID":"749e575d-e292-4ae6-9797-a0af7a566147","Type":"ContainerDied","Data":"450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d"} Nov 24 08:24:57 crc kubenswrapper[4799]: I1124 08:24:57.318351 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qf95" event={"ID":"749e575d-e292-4ae6-9797-a0af7a566147","Type":"ContainerStarted","Data":"bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded"} Nov 24 08:24:57 crc kubenswrapper[4799]: I1124 08:24:57.343078 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5qf95" podStartSLOduration=2.850837299 podStartE2EDuration="6.343055468s" podCreationTimestamp="2025-11-24 08:24:51 +0000 UTC" firstStartedPulling="2025-11-24 08:24:53.2739574 +0000 UTC m=+5838.929939874" lastFinishedPulling="2025-11-24 08:24:56.766175569 +0000 UTC m=+5842.422158043" observedRunningTime="2025-11-24 08:24:57.335952146 +0000 UTC m=+5842.991934640" watchObservedRunningTime="2025-11-24 08:24:57.343055468 +0000 UTC m=+5842.999037942" Nov 24 08:24:57 crc kubenswrapper[4799]: I1124 08:24:57.628375 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:24:57 crc kubenswrapper[4799]: E1124 08:24:57.628964 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:25:01 crc kubenswrapper[4799]: I1124 08:25:01.750971 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:25:01 crc kubenswrapper[4799]: I1124 08:25:01.752453 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:25:01 crc kubenswrapper[4799]: I1124 08:25:01.857939 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:25:02 crc kubenswrapper[4799]: I1124 08:25:02.417257 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:25:02 crc kubenswrapper[4799]: I1124 08:25:02.473937 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5qf95"] Nov 24 08:25:04 crc kubenswrapper[4799]: I1124 08:25:04.381385 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5qf95" podUID="749e575d-e292-4ae6-9797-a0af7a566147" containerName="registry-server" containerID="cri-o://bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded" gracePeriod=2 Nov 24 08:25:04 crc kubenswrapper[4799]: I1124 08:25:04.871321 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:25:04 crc kubenswrapper[4799]: I1124 08:25:04.937701 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjxvz\" (UniqueName: \"kubernetes.io/projected/749e575d-e292-4ae6-9797-a0af7a566147-kube-api-access-mjxvz\") pod \"749e575d-e292-4ae6-9797-a0af7a566147\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " Nov 24 08:25:04 crc kubenswrapper[4799]: I1124 08:25:04.938314 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-catalog-content\") pod \"749e575d-e292-4ae6-9797-a0af7a566147\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " Nov 24 08:25:04 crc kubenswrapper[4799]: I1124 08:25:04.938363 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-utilities\") pod \"749e575d-e292-4ae6-9797-a0af7a566147\" (UID: \"749e575d-e292-4ae6-9797-a0af7a566147\") " Nov 24 08:25:04 crc kubenswrapper[4799]: I1124 08:25:04.939198 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-utilities" (OuterVolumeSpecName: "utilities") pod "749e575d-e292-4ae6-9797-a0af7a566147" (UID: "749e575d-e292-4ae6-9797-a0af7a566147"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:25:04 crc kubenswrapper[4799]: I1124 08:25:04.944646 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/749e575d-e292-4ae6-9797-a0af7a566147-kube-api-access-mjxvz" (OuterVolumeSpecName: "kube-api-access-mjxvz") pod "749e575d-e292-4ae6-9797-a0af7a566147" (UID: "749e575d-e292-4ae6-9797-a0af7a566147"). InnerVolumeSpecName "kube-api-access-mjxvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.005743 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "749e575d-e292-4ae6-9797-a0af7a566147" (UID: "749e575d-e292-4ae6-9797-a0af7a566147"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.040982 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.041022 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjxvz\" (UniqueName: \"kubernetes.io/projected/749e575d-e292-4ae6-9797-a0af7a566147-kube-api-access-mjxvz\") on node \"crc\" DevicePath \"\"" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.041034 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/749e575d-e292-4ae6-9797-a0af7a566147-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.393227 4799 generic.go:334] "Generic (PLEG): container finished" podID="749e575d-e292-4ae6-9797-a0af7a566147" containerID="bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded" exitCode=0 Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.393312 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qf95" event={"ID":"749e575d-e292-4ae6-9797-a0af7a566147","Type":"ContainerDied","Data":"bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded"} Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.394006 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qf95" event={"ID":"749e575d-e292-4ae6-9797-a0af7a566147","Type":"ContainerDied","Data":"56ebe70791a4bba1eb3b1267a3591e2a406ce31ad71763d7fdbc6a14ed245b84"} Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.393354 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qf95" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.394037 4799 scope.go:117] "RemoveContainer" containerID="bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.424564 4799 scope.go:117] "RemoveContainer" containerID="450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.437185 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5qf95"] Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.446262 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5qf95"] Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.465061 4799 scope.go:117] "RemoveContainer" containerID="5c2907b0d63822cbeb2a3fab4d00ab6e475c4ec296a1b2b4ea35f40194279ee8" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.487864 4799 scope.go:117] "RemoveContainer" containerID="bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded" Nov 24 08:25:05 crc kubenswrapper[4799]: E1124 08:25:05.488270 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded\": container with ID starting with bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded not found: ID does not exist" containerID="bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.488308 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded"} err="failed to get container status \"bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded\": rpc error: code = NotFound desc = could not find container \"bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded\": container with ID starting with bbe18df4dc73656e048207d8cc5efe9648c8c328c15a458a67789fb7cdf3cded not found: ID does not exist" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.488330 4799 scope.go:117] "RemoveContainer" containerID="450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d" Nov 24 08:25:05 crc kubenswrapper[4799]: E1124 08:25:05.489567 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d\": container with ID starting with 450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d not found: ID does not exist" containerID="450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.489605 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d"} err="failed to get container status \"450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d\": rpc error: code = NotFound desc = could not find container \"450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d\": container with ID starting with 450ae188e4e529de906ac8052f043a2e731f5956429d2daf17ef4b8cf2b0824d not found: ID does not exist" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.489620 4799 scope.go:117] "RemoveContainer" containerID="5c2907b0d63822cbeb2a3fab4d00ab6e475c4ec296a1b2b4ea35f40194279ee8" Nov 24 08:25:05 crc kubenswrapper[4799]: E1124 08:25:05.489914 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c2907b0d63822cbeb2a3fab4d00ab6e475c4ec296a1b2b4ea35f40194279ee8\": container with ID starting with 5c2907b0d63822cbeb2a3fab4d00ab6e475c4ec296a1b2b4ea35f40194279ee8 not found: ID does not exist" containerID="5c2907b0d63822cbeb2a3fab4d00ab6e475c4ec296a1b2b4ea35f40194279ee8" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.489965 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c2907b0d63822cbeb2a3fab4d00ab6e475c4ec296a1b2b4ea35f40194279ee8"} err="failed to get container status \"5c2907b0d63822cbeb2a3fab4d00ab6e475c4ec296a1b2b4ea35f40194279ee8\": rpc error: code = NotFound desc = could not find container \"5c2907b0d63822cbeb2a3fab4d00ab6e475c4ec296a1b2b4ea35f40194279ee8\": container with ID starting with 5c2907b0d63822cbeb2a3fab4d00ab6e475c4ec296a1b2b4ea35f40194279ee8 not found: ID does not exist" Nov 24 08:25:05 crc kubenswrapper[4799]: I1124 08:25:05.639896 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="749e575d-e292-4ae6-9797-a0af7a566147" path="/var/lib/kubelet/pods/749e575d-e292-4ae6-9797-a0af7a566147/volumes" Nov 24 08:25:10 crc kubenswrapper[4799]: I1124 08:25:10.628124 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:25:10 crc kubenswrapper[4799]: E1124 08:25:10.628984 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:25:10 crc kubenswrapper[4799]: I1124 08:25:10.982365 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cjflv"] Nov 24 08:25:10 crc kubenswrapper[4799]: E1124 08:25:10.982837 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="749e575d-e292-4ae6-9797-a0af7a566147" containerName="extract-content" Nov 24 08:25:10 crc kubenswrapper[4799]: I1124 08:25:10.982878 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="749e575d-e292-4ae6-9797-a0af7a566147" containerName="extract-content" Nov 24 08:25:10 crc kubenswrapper[4799]: E1124 08:25:10.982892 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="749e575d-e292-4ae6-9797-a0af7a566147" containerName="registry-server" Nov 24 08:25:10 crc kubenswrapper[4799]: I1124 08:25:10.982900 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="749e575d-e292-4ae6-9797-a0af7a566147" containerName="registry-server" Nov 24 08:25:10 crc kubenswrapper[4799]: E1124 08:25:10.982929 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="749e575d-e292-4ae6-9797-a0af7a566147" containerName="extract-utilities" Nov 24 08:25:10 crc kubenswrapper[4799]: I1124 08:25:10.982938 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="749e575d-e292-4ae6-9797-a0af7a566147" containerName="extract-utilities" Nov 24 08:25:10 crc kubenswrapper[4799]: I1124 08:25:10.983201 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="749e575d-e292-4ae6-9797-a0af7a566147" containerName="registry-server" Nov 24 08:25:10 crc kubenswrapper[4799]: I1124 08:25:10.984978 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:10 crc kubenswrapper[4799]: I1124 08:25:10.993431 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cjflv"] Nov 24 08:25:11 crc kubenswrapper[4799]: I1124 08:25:11.062421 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-catalog-content\") pod \"redhat-marketplace-cjflv\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:11 crc kubenswrapper[4799]: I1124 08:25:11.062498 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6z47\" (UniqueName: \"kubernetes.io/projected/66609c3c-1faf-486a-acc6-89f4bab7b984-kube-api-access-l6z47\") pod \"redhat-marketplace-cjflv\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:11 crc kubenswrapper[4799]: I1124 08:25:11.062518 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-utilities\") pod \"redhat-marketplace-cjflv\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:11 crc kubenswrapper[4799]: I1124 08:25:11.165015 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-catalog-content\") pod \"redhat-marketplace-cjflv\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:11 crc kubenswrapper[4799]: I1124 08:25:11.165333 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6z47\" (UniqueName: \"kubernetes.io/projected/66609c3c-1faf-486a-acc6-89f4bab7b984-kube-api-access-l6z47\") pod \"redhat-marketplace-cjflv\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:11 crc kubenswrapper[4799]: I1124 08:25:11.165424 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-utilities\") pod \"redhat-marketplace-cjflv\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:11 crc kubenswrapper[4799]: I1124 08:25:11.165819 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-utilities\") pod \"redhat-marketplace-cjflv\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:11 crc kubenswrapper[4799]: I1124 08:25:11.166095 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-catalog-content\") pod \"redhat-marketplace-cjflv\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:11 crc kubenswrapper[4799]: I1124 08:25:11.184802 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6z47\" (UniqueName: \"kubernetes.io/projected/66609c3c-1faf-486a-acc6-89f4bab7b984-kube-api-access-l6z47\") pod \"redhat-marketplace-cjflv\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:11 crc kubenswrapper[4799]: I1124 08:25:11.309196 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:11 crc kubenswrapper[4799]: I1124 08:25:11.795681 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cjflv"] Nov 24 08:25:12 crc kubenswrapper[4799]: I1124 08:25:12.462865 4799 generic.go:334] "Generic (PLEG): container finished" podID="66609c3c-1faf-486a-acc6-89f4bab7b984" containerID="2097ea0e36963a77efe5883104e8540c4dc44dbfd1e16e95682ddfc2df28008e" exitCode=0 Nov 24 08:25:12 crc kubenswrapper[4799]: I1124 08:25:12.462987 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjflv" event={"ID":"66609c3c-1faf-486a-acc6-89f4bab7b984","Type":"ContainerDied","Data":"2097ea0e36963a77efe5883104e8540c4dc44dbfd1e16e95682ddfc2df28008e"} Nov 24 08:25:12 crc kubenswrapper[4799]: I1124 08:25:12.463267 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjflv" event={"ID":"66609c3c-1faf-486a-acc6-89f4bab7b984","Type":"ContainerStarted","Data":"69c0a1afa859dac95dadca7d73c53ff52f8d40023a64bea909ce102ce6baf97f"} Nov 24 08:25:13 crc kubenswrapper[4799]: I1124 08:25:13.472636 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjflv" event={"ID":"66609c3c-1faf-486a-acc6-89f4bab7b984","Type":"ContainerStarted","Data":"c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f"} Nov 24 08:25:14 crc kubenswrapper[4799]: I1124 08:25:14.484719 4799 generic.go:334] "Generic (PLEG): container finished" podID="66609c3c-1faf-486a-acc6-89f4bab7b984" containerID="c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f" exitCode=0 Nov 24 08:25:14 crc kubenswrapper[4799]: I1124 08:25:14.484916 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjflv" event={"ID":"66609c3c-1faf-486a-acc6-89f4bab7b984","Type":"ContainerDied","Data":"c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f"} Nov 24 08:25:15 crc kubenswrapper[4799]: I1124 08:25:15.494162 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjflv" event={"ID":"66609c3c-1faf-486a-acc6-89f4bab7b984","Type":"ContainerStarted","Data":"74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56"} Nov 24 08:25:15 crc kubenswrapper[4799]: I1124 08:25:15.515609 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cjflv" podStartSLOduration=2.910444508 podStartE2EDuration="5.515585838s" podCreationTimestamp="2025-11-24 08:25:10 +0000 UTC" firstStartedPulling="2025-11-24 08:25:12.466021413 +0000 UTC m=+5858.122003907" lastFinishedPulling="2025-11-24 08:25:15.071162763 +0000 UTC m=+5860.727145237" observedRunningTime="2025-11-24 08:25:15.514652572 +0000 UTC m=+5861.170635056" watchObservedRunningTime="2025-11-24 08:25:15.515585838 +0000 UTC m=+5861.171568332" Nov 24 08:25:21 crc kubenswrapper[4799]: I1124 08:25:21.310273 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:21 crc kubenswrapper[4799]: I1124 08:25:21.311490 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:21 crc kubenswrapper[4799]: I1124 08:25:21.367533 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:21 crc kubenswrapper[4799]: I1124 08:25:21.632959 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:25:21 crc kubenswrapper[4799]: E1124 08:25:21.633345 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:25:21 crc kubenswrapper[4799]: I1124 08:25:21.641233 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:22 crc kubenswrapper[4799]: I1124 08:25:22.606934 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cjflv"] Nov 24 08:25:24 crc kubenswrapper[4799]: I1124 08:25:24.596104 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cjflv" podUID="66609c3c-1faf-486a-acc6-89f4bab7b984" containerName="registry-server" containerID="cri-o://74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56" gracePeriod=2 Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.136265 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.271943 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-utilities\") pod \"66609c3c-1faf-486a-acc6-89f4bab7b984\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.272516 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-catalog-content\") pod \"66609c3c-1faf-486a-acc6-89f4bab7b984\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.272640 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6z47\" (UniqueName: \"kubernetes.io/projected/66609c3c-1faf-486a-acc6-89f4bab7b984-kube-api-access-l6z47\") pod \"66609c3c-1faf-486a-acc6-89f4bab7b984\" (UID: \"66609c3c-1faf-486a-acc6-89f4bab7b984\") " Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.273561 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-utilities" (OuterVolumeSpecName: "utilities") pod "66609c3c-1faf-486a-acc6-89f4bab7b984" (UID: "66609c3c-1faf-486a-acc6-89f4bab7b984"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.280176 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66609c3c-1faf-486a-acc6-89f4bab7b984-kube-api-access-l6z47" (OuterVolumeSpecName: "kube-api-access-l6z47") pod "66609c3c-1faf-486a-acc6-89f4bab7b984" (UID: "66609c3c-1faf-486a-acc6-89f4bab7b984"). InnerVolumeSpecName "kube-api-access-l6z47". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.300710 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66609c3c-1faf-486a-acc6-89f4bab7b984" (UID: "66609c3c-1faf-486a-acc6-89f4bab7b984"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.375264 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.375302 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6z47\" (UniqueName: \"kubernetes.io/projected/66609c3c-1faf-486a-acc6-89f4bab7b984-kube-api-access-l6z47\") on node \"crc\" DevicePath \"\"" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.375315 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66609c3c-1faf-486a-acc6-89f4bab7b984-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.627014 4799 generic.go:334] "Generic (PLEG): container finished" podID="66609c3c-1faf-486a-acc6-89f4bab7b984" containerID="74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56" exitCode=0 Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.627082 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjflv" event={"ID":"66609c3c-1faf-486a-acc6-89f4bab7b984","Type":"ContainerDied","Data":"74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56"} Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.627119 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjflv" event={"ID":"66609c3c-1faf-486a-acc6-89f4bab7b984","Type":"ContainerDied","Data":"69c0a1afa859dac95dadca7d73c53ff52f8d40023a64bea909ce102ce6baf97f"} Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.627140 4799 scope.go:117] "RemoveContainer" containerID="74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.627419 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cjflv" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.650185 4799 scope.go:117] "RemoveContainer" containerID="c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.678133 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cjflv"] Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.680616 4799 scope.go:117] "RemoveContainer" containerID="2097ea0e36963a77efe5883104e8540c4dc44dbfd1e16e95682ddfc2df28008e" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.687891 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cjflv"] Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.735206 4799 scope.go:117] "RemoveContainer" containerID="74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56" Nov 24 08:25:25 crc kubenswrapper[4799]: E1124 08:25:25.735786 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56\": container with ID starting with 74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56 not found: ID does not exist" containerID="74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.735831 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56"} err="failed to get container status \"74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56\": rpc error: code = NotFound desc = could not find container \"74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56\": container with ID starting with 74fcb20bfc1b50d3965e94762fdd2fddc2aca9a64964a1848fe7005db15f8d56 not found: ID does not exist" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.735880 4799 scope.go:117] "RemoveContainer" containerID="c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f" Nov 24 08:25:25 crc kubenswrapper[4799]: E1124 08:25:25.736233 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f\": container with ID starting with c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f not found: ID does not exist" containerID="c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.736272 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f"} err="failed to get container status \"c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f\": rpc error: code = NotFound desc = could not find container \"c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f\": container with ID starting with c93f74a163f1833210442558335a4c2a9ea25c224e253ea3eb5b97fa7c5a722f not found: ID does not exist" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.736296 4799 scope.go:117] "RemoveContainer" containerID="2097ea0e36963a77efe5883104e8540c4dc44dbfd1e16e95682ddfc2df28008e" Nov 24 08:25:25 crc kubenswrapper[4799]: E1124 08:25:25.736690 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2097ea0e36963a77efe5883104e8540c4dc44dbfd1e16e95682ddfc2df28008e\": container with ID starting with 2097ea0e36963a77efe5883104e8540c4dc44dbfd1e16e95682ddfc2df28008e not found: ID does not exist" containerID="2097ea0e36963a77efe5883104e8540c4dc44dbfd1e16e95682ddfc2df28008e" Nov 24 08:25:25 crc kubenswrapper[4799]: I1124 08:25:25.736713 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2097ea0e36963a77efe5883104e8540c4dc44dbfd1e16e95682ddfc2df28008e"} err="failed to get container status \"2097ea0e36963a77efe5883104e8540c4dc44dbfd1e16e95682ddfc2df28008e\": rpc error: code = NotFound desc = could not find container \"2097ea0e36963a77efe5883104e8540c4dc44dbfd1e16e95682ddfc2df28008e\": container with ID starting with 2097ea0e36963a77efe5883104e8540c4dc44dbfd1e16e95682ddfc2df28008e not found: ID does not exist" Nov 24 08:25:27 crc kubenswrapper[4799]: I1124 08:25:27.655339 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66609c3c-1faf-486a-acc6-89f4bab7b984" path="/var/lib/kubelet/pods/66609c3c-1faf-486a-acc6-89f4bab7b984/volumes" Nov 24 08:25:32 crc kubenswrapper[4799]: I1124 08:25:32.628494 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:25:32 crc kubenswrapper[4799]: E1124 08:25:32.629547 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:25:45 crc kubenswrapper[4799]: I1124 08:25:45.645642 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:25:45 crc kubenswrapper[4799]: E1124 08:25:45.646436 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:25:51 crc kubenswrapper[4799]: I1124 08:25:51.043701 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-pzp8c"] Nov 24 08:25:51 crc kubenswrapper[4799]: I1124 08:25:51.052781 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-pzp8c"] Nov 24 08:25:51 crc kubenswrapper[4799]: I1124 08:25:51.653642 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bb0216b-56a7-4822-bfce-0cf463a9d595" path="/var/lib/kubelet/pods/5bb0216b-56a7-4822-bfce-0cf463a9d595/volumes" Nov 24 08:25:52 crc kubenswrapper[4799]: I1124 08:25:52.031623 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-7167-account-create-nxvxk"] Nov 24 08:25:52 crc kubenswrapper[4799]: I1124 08:25:52.039805 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-7167-account-create-nxvxk"] Nov 24 08:25:53 crc kubenswrapper[4799]: I1124 08:25:53.641063 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="388f8341-6e2d-40e3-b35c-d1bd0d95e134" path="/var/lib/kubelet/pods/388f8341-6e2d-40e3-b35c-d1bd0d95e134/volumes" Nov 24 08:25:58 crc kubenswrapper[4799]: I1124 08:25:58.628666 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:25:58 crc kubenswrapper[4799]: E1124 08:25:58.629875 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:26:07 crc kubenswrapper[4799]: I1124 08:26:07.041861 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-7bnw6"] Nov 24 08:26:07 crc kubenswrapper[4799]: I1124 08:26:07.049230 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-7bnw6"] Nov 24 08:26:07 crc kubenswrapper[4799]: I1124 08:26:07.644132 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6acabb45-e7c1-447f-aaaa-2a3514e3a80d" path="/var/lib/kubelet/pods/6acabb45-e7c1-447f-aaaa-2a3514e3a80d/volumes" Nov 24 08:26:10 crc kubenswrapper[4799]: I1124 08:26:10.198087 4799 scope.go:117] "RemoveContainer" containerID="d3abb67370ec878d7c0d6f808292579923792f83eb0f919b804aac968f4ee28a" Nov 24 08:26:10 crc kubenswrapper[4799]: I1124 08:26:10.245597 4799 scope.go:117] "RemoveContainer" containerID="b6510b4c8324aeec824855d1bef9711d745c4301afa6e642fa9998ee573fcc09" Nov 24 08:26:10 crc kubenswrapper[4799]: I1124 08:26:10.328643 4799 scope.go:117] "RemoveContainer" containerID="f3b03d3bdd29c910243b0c4ab8146de727a5be380550f3c756f2cf63400c6576" Nov 24 08:26:13 crc kubenswrapper[4799]: I1124 08:26:13.629380 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:26:13 crc kubenswrapper[4799]: E1124 08:26:13.630113 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:26:25 crc kubenswrapper[4799]: I1124 08:26:25.634974 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:26:25 crc kubenswrapper[4799]: E1124 08:26:25.636174 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:26:38 crc kubenswrapper[4799]: I1124 08:26:38.628221 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:26:38 crc kubenswrapper[4799]: E1124 08:26:38.629404 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:26:53 crc kubenswrapper[4799]: I1124 08:26:53.630327 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:26:54 crc kubenswrapper[4799]: I1124 08:26:54.598935 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"24867b1a1093471fa92100e6059033c95aa57e48ac37906d109329e71b7be940"} Nov 24 08:27:10 crc kubenswrapper[4799]: I1124 08:27:10.465693 4799 scope.go:117] "RemoveContainer" containerID="4eb655cf30825d2c546aae9992cad1e4f2a79bbd5fa14cd04efc789ca6f86272" Nov 24 08:27:10 crc kubenswrapper[4799]: I1124 08:27:10.503392 4799 scope.go:117] "RemoveContainer" containerID="1fd1259207e22fb320c451513f0237651bd41e2208d835dda8a2918955ec64c2" Nov 24 08:28:20 crc kubenswrapper[4799]: I1124 08:28:20.044298 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-ec59-account-create-8nmkj"] Nov 24 08:28:20 crc kubenswrapper[4799]: I1124 08:28:20.053598 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-ec59-account-create-8nmkj"] Nov 24 08:28:20 crc kubenswrapper[4799]: I1124 08:28:20.080407 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-w4dc6"] Nov 24 08:28:20 crc kubenswrapper[4799]: I1124 08:28:20.084783 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-w4dc6"] Nov 24 08:28:21 crc kubenswrapper[4799]: I1124 08:28:21.642084 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31932a1e-5342-4097-8e42-8855cf672708" path="/var/lib/kubelet/pods/31932a1e-5342-4097-8e42-8855cf672708/volumes" Nov 24 08:28:21 crc kubenswrapper[4799]: I1124 08:28:21.643166 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d87dc8df-3799-495f-8090-536332563977" path="/var/lib/kubelet/pods/d87dc8df-3799-495f-8090-536332563977/volumes" Nov 24 08:28:32 crc kubenswrapper[4799]: I1124 08:28:32.039641 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-wvq24"] Nov 24 08:28:32 crc kubenswrapper[4799]: I1124 08:28:32.046745 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-wvq24"] Nov 24 08:28:33 crc kubenswrapper[4799]: I1124 08:28:33.656818 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81" path="/var/lib/kubelet/pods/80c7c70c-a3fe-46ab-b48f-a3b8cf5e3d81/volumes" Nov 24 08:28:52 crc kubenswrapper[4799]: I1124 08:28:52.036512 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-55dd-account-create-x7dhk"] Nov 24 08:28:52 crc kubenswrapper[4799]: I1124 08:28:52.047009 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-j5n76"] Nov 24 08:28:52 crc kubenswrapper[4799]: I1124 08:28:52.055322 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-55dd-account-create-x7dhk"] Nov 24 08:28:52 crc kubenswrapper[4799]: I1124 08:28:52.063197 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-j5n76"] Nov 24 08:28:53 crc kubenswrapper[4799]: I1124 08:28:53.643571 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95e37915-feda-469a-bcb5-983988c362e5" path="/var/lib/kubelet/pods/95e37915-feda-469a-bcb5-983988c362e5/volumes" Nov 24 08:28:53 crc kubenswrapper[4799]: I1124 08:28:53.646525 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1025848-a743-477d-9309-0b60871a2025" path="/var/lib/kubelet/pods/f1025848-a743-477d-9309-0b60871a2025/volumes" Nov 24 08:29:05 crc kubenswrapper[4799]: I1124 08:29:05.041788 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-9jmx6"] Nov 24 08:29:05 crc kubenswrapper[4799]: I1124 08:29:05.051224 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-9jmx6"] Nov 24 08:29:05 crc kubenswrapper[4799]: I1124 08:29:05.640414 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76125ed1-4d2e-4fe8-9f73-c66425bd8506" path="/var/lib/kubelet/pods/76125ed1-4d2e-4fe8-9f73-c66425bd8506/volumes" Nov 24 08:29:10 crc kubenswrapper[4799]: I1124 08:29:10.597229 4799 scope.go:117] "RemoveContainer" containerID="dce6e19b87b53b1c51b330203a18e4aa042fc14397acf2e0e92ffe3d992713a1" Nov 24 08:29:10 crc kubenswrapper[4799]: I1124 08:29:10.633540 4799 scope.go:117] "RemoveContainer" containerID="05dd68c42b18d98a47730e3a26ba6fbef0b517be3da5fea3622f62acb45c1bd4" Nov 24 08:29:10 crc kubenswrapper[4799]: I1124 08:29:10.696139 4799 scope.go:117] "RemoveContainer" containerID="6db2b871502bd4c16168a3adc2db8d87a64ab75a32d7901a1135f05dbe9bf1fd" Nov 24 08:29:10 crc kubenswrapper[4799]: I1124 08:29:10.748729 4799 scope.go:117] "RemoveContainer" containerID="562e27c5270effa8f4a9ab891eb34f5ca34b1515e27a9117556cc4ea7dc0e67e" Nov 24 08:29:10 crc kubenswrapper[4799]: I1124 08:29:10.774780 4799 scope.go:117] "RemoveContainer" containerID="5934d736be852a953e192a06ec301201db426ef57fdfcd965b85635527c406e7" Nov 24 08:29:10 crc kubenswrapper[4799]: I1124 08:29:10.821273 4799 scope.go:117] "RemoveContainer" containerID="bc522741670ba4c2268e6b10472e01d059aa435d62902403a626d3477676aa2b" Nov 24 08:29:20 crc kubenswrapper[4799]: I1124 08:29:20.400289 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:29:20 crc kubenswrapper[4799]: I1124 08:29:20.400908 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:29:50 crc kubenswrapper[4799]: I1124 08:29:50.400568 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:29:50 crc kubenswrapper[4799]: I1124 08:29:50.401160 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.145058 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt"] Nov 24 08:30:00 crc kubenswrapper[4799]: E1124 08:30:00.146167 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66609c3c-1faf-486a-acc6-89f4bab7b984" containerName="registry-server" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.146183 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="66609c3c-1faf-486a-acc6-89f4bab7b984" containerName="registry-server" Nov 24 08:30:00 crc kubenswrapper[4799]: E1124 08:30:00.146199 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66609c3c-1faf-486a-acc6-89f4bab7b984" containerName="extract-content" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.146204 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="66609c3c-1faf-486a-acc6-89f4bab7b984" containerName="extract-content" Nov 24 08:30:00 crc kubenswrapper[4799]: E1124 08:30:00.146216 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66609c3c-1faf-486a-acc6-89f4bab7b984" containerName="extract-utilities" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.146222 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="66609c3c-1faf-486a-acc6-89f4bab7b984" containerName="extract-utilities" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.146441 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="66609c3c-1faf-486a-acc6-89f4bab7b984" containerName="registry-server" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.147299 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.150161 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.150417 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.158353 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt"] Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.228707 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8321770-f41c-4952-a7d0-a8fa94543a22-config-volume\") pod \"collect-profiles-29399550-8vztt\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.228909 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgzvb\" (UniqueName: \"kubernetes.io/projected/e8321770-f41c-4952-a7d0-a8fa94543a22-kube-api-access-pgzvb\") pod \"collect-profiles-29399550-8vztt\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.228950 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8321770-f41c-4952-a7d0-a8fa94543a22-secret-volume\") pod \"collect-profiles-29399550-8vztt\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.331541 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8321770-f41c-4952-a7d0-a8fa94543a22-config-volume\") pod \"collect-profiles-29399550-8vztt\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.331657 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgzvb\" (UniqueName: \"kubernetes.io/projected/e8321770-f41c-4952-a7d0-a8fa94543a22-kube-api-access-pgzvb\") pod \"collect-profiles-29399550-8vztt\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.331705 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8321770-f41c-4952-a7d0-a8fa94543a22-secret-volume\") pod \"collect-profiles-29399550-8vztt\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.332507 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8321770-f41c-4952-a7d0-a8fa94543a22-config-volume\") pod \"collect-profiles-29399550-8vztt\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.338224 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8321770-f41c-4952-a7d0-a8fa94543a22-secret-volume\") pod \"collect-profiles-29399550-8vztt\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.364073 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgzvb\" (UniqueName: \"kubernetes.io/projected/e8321770-f41c-4952-a7d0-a8fa94543a22-kube-api-access-pgzvb\") pod \"collect-profiles-29399550-8vztt\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:00 crc kubenswrapper[4799]: I1124 08:30:00.473284 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:01 crc kubenswrapper[4799]: I1124 08:30:00.927448 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt"] Nov 24 08:30:01 crc kubenswrapper[4799]: I1124 08:30:01.508083 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" event={"ID":"e8321770-f41c-4952-a7d0-a8fa94543a22","Type":"ContainerStarted","Data":"b64a918bcf5533d6068009fa3af7fdce739558277ccf7a90942505a32f7afe2a"} Nov 24 08:30:01 crc kubenswrapper[4799]: I1124 08:30:01.508429 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" event={"ID":"e8321770-f41c-4952-a7d0-a8fa94543a22","Type":"ContainerStarted","Data":"3e9a8571b103bad2eb7b3cdd0e39dd481d0306cd1ee203d52f3301bb9753d793"} Nov 24 08:30:01 crc kubenswrapper[4799]: I1124 08:30:01.525194 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" podStartSLOduration=1.5251777880000001 podStartE2EDuration="1.525177788s" podCreationTimestamp="2025-11-24 08:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:30:01.522194353 +0000 UTC m=+6147.178176827" watchObservedRunningTime="2025-11-24 08:30:01.525177788 +0000 UTC m=+6147.181160262" Nov 24 08:30:02 crc kubenswrapper[4799]: I1124 08:30:02.520491 4799 generic.go:334] "Generic (PLEG): container finished" podID="e8321770-f41c-4952-a7d0-a8fa94543a22" containerID="b64a918bcf5533d6068009fa3af7fdce739558277ccf7a90942505a32f7afe2a" exitCode=0 Nov 24 08:30:02 crc kubenswrapper[4799]: I1124 08:30:02.520594 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" event={"ID":"e8321770-f41c-4952-a7d0-a8fa94543a22","Type":"ContainerDied","Data":"b64a918bcf5533d6068009fa3af7fdce739558277ccf7a90942505a32f7afe2a"} Nov 24 08:30:03 crc kubenswrapper[4799]: I1124 08:30:03.912058 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.000246 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8321770-f41c-4952-a7d0-a8fa94543a22-config-volume\") pod \"e8321770-f41c-4952-a7d0-a8fa94543a22\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.000390 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8321770-f41c-4952-a7d0-a8fa94543a22-secret-volume\") pod \"e8321770-f41c-4952-a7d0-a8fa94543a22\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.000519 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgzvb\" (UniqueName: \"kubernetes.io/projected/e8321770-f41c-4952-a7d0-a8fa94543a22-kube-api-access-pgzvb\") pod \"e8321770-f41c-4952-a7d0-a8fa94543a22\" (UID: \"e8321770-f41c-4952-a7d0-a8fa94543a22\") " Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.002512 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8321770-f41c-4952-a7d0-a8fa94543a22-config-volume" (OuterVolumeSpecName: "config-volume") pod "e8321770-f41c-4952-a7d0-a8fa94543a22" (UID: "e8321770-f41c-4952-a7d0-a8fa94543a22"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.006663 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8321770-f41c-4952-a7d0-a8fa94543a22-kube-api-access-pgzvb" (OuterVolumeSpecName: "kube-api-access-pgzvb") pod "e8321770-f41c-4952-a7d0-a8fa94543a22" (UID: "e8321770-f41c-4952-a7d0-a8fa94543a22"). InnerVolumeSpecName "kube-api-access-pgzvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.011114 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8321770-f41c-4952-a7d0-a8fa94543a22-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e8321770-f41c-4952-a7d0-a8fa94543a22" (UID: "e8321770-f41c-4952-a7d0-a8fa94543a22"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.103220 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgzvb\" (UniqueName: \"kubernetes.io/projected/e8321770-f41c-4952-a7d0-a8fa94543a22-kube-api-access-pgzvb\") on node \"crc\" DevicePath \"\"" Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.103255 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8321770-f41c-4952-a7d0-a8fa94543a22-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.103269 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8321770-f41c-4952-a7d0-a8fa94543a22-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.549206 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" event={"ID":"e8321770-f41c-4952-a7d0-a8fa94543a22","Type":"ContainerDied","Data":"3e9a8571b103bad2eb7b3cdd0e39dd481d0306cd1ee203d52f3301bb9753d793"} Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.549580 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e9a8571b103bad2eb7b3cdd0e39dd481d0306cd1ee203d52f3301bb9753d793" Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.549418 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt" Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.590331 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb"] Nov 24 08:30:04 crc kubenswrapper[4799]: I1124 08:30:04.597619 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399505-b2gvb"] Nov 24 08:30:05 crc kubenswrapper[4799]: I1124 08:30:05.644996 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2772baa3-685c-40ae-b995-757ee71a5fcd" path="/var/lib/kubelet/pods/2772baa3-685c-40ae-b995-757ee71a5fcd/volumes" Nov 24 08:30:11 crc kubenswrapper[4799]: I1124 08:30:11.083665 4799 scope.go:117] "RemoveContainer" containerID="7f870487cd6f8acf504f8baca030981897011639e9ddd4b22dd7dcf6e30952d5" Nov 24 08:30:20 crc kubenswrapper[4799]: I1124 08:30:20.400401 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:30:20 crc kubenswrapper[4799]: I1124 08:30:20.401008 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:30:20 crc kubenswrapper[4799]: I1124 08:30:20.401062 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 08:30:20 crc kubenswrapper[4799]: I1124 08:30:20.401967 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"24867b1a1093471fa92100e6059033c95aa57e48ac37906d109329e71b7be940"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:30:20 crc kubenswrapper[4799]: I1124 08:30:20.402033 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://24867b1a1093471fa92100e6059033c95aa57e48ac37906d109329e71b7be940" gracePeriod=600 Nov 24 08:30:20 crc kubenswrapper[4799]: I1124 08:30:20.699503 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="24867b1a1093471fa92100e6059033c95aa57e48ac37906d109329e71b7be940" exitCode=0 Nov 24 08:30:20 crc kubenswrapper[4799]: I1124 08:30:20.699552 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"24867b1a1093471fa92100e6059033c95aa57e48ac37906d109329e71b7be940"} Nov 24 08:30:20 crc kubenswrapper[4799]: I1124 08:30:20.699603 4799 scope.go:117] "RemoveContainer" containerID="2517cd495c9ed4328a760340aa2f186761f5f42592ca5e9374d44b4978cc3c72" Nov 24 08:30:21 crc kubenswrapper[4799]: I1124 08:30:21.711378 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b"} Nov 24 08:31:00 crc kubenswrapper[4799]: I1124 08:31:00.113234 4799 generic.go:334] "Generic (PLEG): container finished" podID="23397601-e699-4796-804a-bcf6b64f4a20" containerID="e750eb4dc8ddab1e5378123bdf61a1bae1c120359c4efd2311d3f7e1b049a80c" exitCode=0 Nov 24 08:31:00 crc kubenswrapper[4799]: I1124 08:31:00.113326 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" event={"ID":"23397601-e699-4796-804a-bcf6b64f4a20","Type":"ContainerDied","Data":"e750eb4dc8ddab1e5378123bdf61a1bae1c120359c4efd2311d3f7e1b049a80c"} Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.585208 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.725168 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ceph\") pod \"23397601-e699-4796-804a-bcf6b64f4a20\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.725228 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-tripleo-cleanup-combined-ca-bundle\") pod \"23397601-e699-4796-804a-bcf6b64f4a20\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.725337 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ssh-key\") pod \"23397601-e699-4796-804a-bcf6b64f4a20\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.725385 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-inventory\") pod \"23397601-e699-4796-804a-bcf6b64f4a20\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.725444 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dszmf\" (UniqueName: \"kubernetes.io/projected/23397601-e699-4796-804a-bcf6b64f4a20-kube-api-access-dszmf\") pod \"23397601-e699-4796-804a-bcf6b64f4a20\" (UID: \"23397601-e699-4796-804a-bcf6b64f4a20\") " Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.732491 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23397601-e699-4796-804a-bcf6b64f4a20-kube-api-access-dszmf" (OuterVolumeSpecName: "kube-api-access-dszmf") pod "23397601-e699-4796-804a-bcf6b64f4a20" (UID: "23397601-e699-4796-804a-bcf6b64f4a20"). InnerVolumeSpecName "kube-api-access-dszmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.732696 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ceph" (OuterVolumeSpecName: "ceph") pod "23397601-e699-4796-804a-bcf6b64f4a20" (UID: "23397601-e699-4796-804a-bcf6b64f4a20"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.736326 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "23397601-e699-4796-804a-bcf6b64f4a20" (UID: "23397601-e699-4796-804a-bcf6b64f4a20"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.757081 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "23397601-e699-4796-804a-bcf6b64f4a20" (UID: "23397601-e699-4796-804a-bcf6b64f4a20"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.766371 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-inventory" (OuterVolumeSpecName: "inventory") pod "23397601-e699-4796-804a-bcf6b64f4a20" (UID: "23397601-e699-4796-804a-bcf6b64f4a20"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.827700 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.827732 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.827742 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dszmf\" (UniqueName: \"kubernetes.io/projected/23397601-e699-4796-804a-bcf6b64f4a20-kube-api-access-dszmf\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.827752 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:01 crc kubenswrapper[4799]: I1124 08:31:01.827762 4799 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23397601-e699-4796-804a-bcf6b64f4a20-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:02 crc kubenswrapper[4799]: I1124 08:31:02.141468 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" event={"ID":"23397601-e699-4796-804a-bcf6b64f4a20","Type":"ContainerDied","Data":"6dab5a0ae89040e6fec768ad0589646fe8373b08f8114ce123c45ff0a20ceef9"} Nov 24 08:31:02 crc kubenswrapper[4799]: I1124 08:31:02.141549 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dab5a0ae89040e6fec768ad0589646fe8373b08f8114ce123c45ff0a20ceef9" Nov 24 08:31:02 crc kubenswrapper[4799]: I1124 08:31:02.141688 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.546564 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-lkqh5"] Nov 24 08:31:05 crc kubenswrapper[4799]: E1124 08:31:05.547686 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23397601-e699-4796-804a-bcf6b64f4a20" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.547704 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="23397601-e699-4796-804a-bcf6b64f4a20" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 24 08:31:05 crc kubenswrapper[4799]: E1124 08:31:05.547725 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8321770-f41c-4952-a7d0-a8fa94543a22" containerName="collect-profiles" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.547733 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8321770-f41c-4952-a7d0-a8fa94543a22" containerName="collect-profiles" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.548006 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8321770-f41c-4952-a7d0-a8fa94543a22" containerName="collect-profiles" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.548042 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="23397601-e699-4796-804a-bcf6b64f4a20" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.548948 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.552144 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.552203 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.552895 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.553370 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.555578 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-lkqh5"] Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.613466 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.613532 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2bs9\" (UniqueName: \"kubernetes.io/projected/fefc2996-4fc3-40b2-bf36-82c82ba5039a-kube-api-access-t2bs9\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.613570 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-inventory\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.614085 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.614365 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ceph\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.715566 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2bs9\" (UniqueName: \"kubernetes.io/projected/fefc2996-4fc3-40b2-bf36-82c82ba5039a-kube-api-access-t2bs9\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.715617 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-inventory\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.715756 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.715813 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ceph\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.715885 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.723029 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.723229 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ceph\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.723434 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.725250 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-inventory\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.734251 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2bs9\" (UniqueName: \"kubernetes.io/projected/fefc2996-4fc3-40b2-bf36-82c82ba5039a-kube-api-access-t2bs9\") pod \"bootstrap-openstack-openstack-cell1-lkqh5\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:05 crc kubenswrapper[4799]: I1124 08:31:05.873061 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:31:06 crc kubenswrapper[4799]: I1124 08:31:06.496334 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-lkqh5"] Nov 24 08:31:06 crc kubenswrapper[4799]: I1124 08:31:06.504185 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:31:07 crc kubenswrapper[4799]: I1124 08:31:07.200992 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" event={"ID":"fefc2996-4fc3-40b2-bf36-82c82ba5039a","Type":"ContainerStarted","Data":"8b75a48bd7e9e3b8f7aa8f22f1ed8af702d6f259e1d307ac955eaa704984fd9e"} Nov 24 08:31:08 crc kubenswrapper[4799]: I1124 08:31:08.215918 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" event={"ID":"fefc2996-4fc3-40b2-bf36-82c82ba5039a","Type":"ContainerStarted","Data":"49d33ad0efcdb1ef4b49b2c386b5b14a39aa96031ee6daa726efeec118cef766"} Nov 24 08:31:08 crc kubenswrapper[4799]: I1124 08:31:08.248138 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" podStartSLOduration=2.0794322689999998 podStartE2EDuration="3.248113779s" podCreationTimestamp="2025-11-24 08:31:05 +0000 UTC" firstStartedPulling="2025-11-24 08:31:06.5039684 +0000 UTC m=+6212.159950874" lastFinishedPulling="2025-11-24 08:31:07.67264988 +0000 UTC m=+6213.328632384" observedRunningTime="2025-11-24 08:31:08.238707592 +0000 UTC m=+6213.894690076" watchObservedRunningTime="2025-11-24 08:31:08.248113779 +0000 UTC m=+6213.904096263" Nov 24 08:32:20 crc kubenswrapper[4799]: I1124 08:32:20.400823 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:32:20 crc kubenswrapper[4799]: I1124 08:32:20.401589 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:32:50 crc kubenswrapper[4799]: I1124 08:32:50.400946 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:32:50 crc kubenswrapper[4799]: I1124 08:32:50.402097 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:33:20 crc kubenswrapper[4799]: I1124 08:33:20.400964 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:33:20 crc kubenswrapper[4799]: I1124 08:33:20.401777 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:33:20 crc kubenswrapper[4799]: I1124 08:33:20.401886 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 08:33:20 crc kubenswrapper[4799]: I1124 08:33:20.403630 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:33:20 crc kubenswrapper[4799]: I1124 08:33:20.403708 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" gracePeriod=600 Nov 24 08:33:20 crc kubenswrapper[4799]: E1124 08:33:20.561909 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:33:20 crc kubenswrapper[4799]: I1124 08:33:20.593317 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" exitCode=0 Nov 24 08:33:20 crc kubenswrapper[4799]: I1124 08:33:20.593389 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b"} Nov 24 08:33:20 crc kubenswrapper[4799]: I1124 08:33:20.593434 4799 scope.go:117] "RemoveContainer" containerID="24867b1a1093471fa92100e6059033c95aa57e48ac37906d109329e71b7be940" Nov 24 08:33:20 crc kubenswrapper[4799]: I1124 08:33:20.594314 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:33:20 crc kubenswrapper[4799]: E1124 08:33:20.594654 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.191417 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f2zpr"] Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.202047 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.220654 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f2zpr"] Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.279503 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-catalog-content\") pod \"certified-operators-f2zpr\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.279680 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-utilities\") pod \"certified-operators-f2zpr\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.279792 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js9zs\" (UniqueName: \"kubernetes.io/projected/b70a08e5-6963-4218-a193-24b52e593f7c-kube-api-access-js9zs\") pod \"certified-operators-f2zpr\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.381690 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js9zs\" (UniqueName: \"kubernetes.io/projected/b70a08e5-6963-4218-a193-24b52e593f7c-kube-api-access-js9zs\") pod \"certified-operators-f2zpr\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.381757 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-catalog-content\") pod \"certified-operators-f2zpr\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.381877 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-utilities\") pod \"certified-operators-f2zpr\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.382388 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-utilities\") pod \"certified-operators-f2zpr\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.383129 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-catalog-content\") pod \"certified-operators-f2zpr\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.408260 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js9zs\" (UniqueName: \"kubernetes.io/projected/b70a08e5-6963-4218-a193-24b52e593f7c-kube-api-access-js9zs\") pod \"certified-operators-f2zpr\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:23 crc kubenswrapper[4799]: I1124 08:33:23.575613 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:24 crc kubenswrapper[4799]: I1124 08:33:24.057172 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f2zpr"] Nov 24 08:33:24 crc kubenswrapper[4799]: I1124 08:33:24.632250 4799 generic.go:334] "Generic (PLEG): container finished" podID="b70a08e5-6963-4218-a193-24b52e593f7c" containerID="f3460dab6b4fb4634ee2c692b9b2a8d8cd3027a02ea6bed1d7c76cc8376311be" exitCode=0 Nov 24 08:33:24 crc kubenswrapper[4799]: I1124 08:33:24.632333 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f2zpr" event={"ID":"b70a08e5-6963-4218-a193-24b52e593f7c","Type":"ContainerDied","Data":"f3460dab6b4fb4634ee2c692b9b2a8d8cd3027a02ea6bed1d7c76cc8376311be"} Nov 24 08:33:24 crc kubenswrapper[4799]: I1124 08:33:24.632530 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f2zpr" event={"ID":"b70a08e5-6963-4218-a193-24b52e593f7c","Type":"ContainerStarted","Data":"2dfb1158c952e8a47d3425e2e36f0bf5158c452a524cfd1df84fd1170eef00f2"} Nov 24 08:33:25 crc kubenswrapper[4799]: I1124 08:33:25.648001 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f2zpr" event={"ID":"b70a08e5-6963-4218-a193-24b52e593f7c","Type":"ContainerStarted","Data":"06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c"} Nov 24 08:33:26 crc kubenswrapper[4799]: I1124 08:33:26.652188 4799 generic.go:334] "Generic (PLEG): container finished" podID="b70a08e5-6963-4218-a193-24b52e593f7c" containerID="06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c" exitCode=0 Nov 24 08:33:26 crc kubenswrapper[4799]: I1124 08:33:26.652553 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f2zpr" event={"ID":"b70a08e5-6963-4218-a193-24b52e593f7c","Type":"ContainerDied","Data":"06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c"} Nov 24 08:33:27 crc kubenswrapper[4799]: I1124 08:33:27.664077 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f2zpr" event={"ID":"b70a08e5-6963-4218-a193-24b52e593f7c","Type":"ContainerStarted","Data":"3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452"} Nov 24 08:33:27 crc kubenswrapper[4799]: I1124 08:33:27.685288 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f2zpr" podStartSLOduration=2.088324207 podStartE2EDuration="4.685269293s" podCreationTimestamp="2025-11-24 08:33:23 +0000 UTC" firstStartedPulling="2025-11-24 08:33:24.635031729 +0000 UTC m=+6350.291014233" lastFinishedPulling="2025-11-24 08:33:27.231976845 +0000 UTC m=+6352.887959319" observedRunningTime="2025-11-24 08:33:27.680463776 +0000 UTC m=+6353.336446290" watchObservedRunningTime="2025-11-24 08:33:27.685269293 +0000 UTC m=+6353.341251767" Nov 24 08:33:31 crc kubenswrapper[4799]: I1124 08:33:31.628755 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:33:31 crc kubenswrapper[4799]: E1124 08:33:31.629815 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:33:33 crc kubenswrapper[4799]: I1124 08:33:33.576496 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:33 crc kubenswrapper[4799]: I1124 08:33:33.576557 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:33 crc kubenswrapper[4799]: I1124 08:33:33.642309 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:33 crc kubenswrapper[4799]: I1124 08:33:33.766688 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:33 crc kubenswrapper[4799]: I1124 08:33:33.889731 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f2zpr"] Nov 24 08:33:35 crc kubenswrapper[4799]: I1124 08:33:35.737216 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f2zpr" podUID="b70a08e5-6963-4218-a193-24b52e593f7c" containerName="registry-server" containerID="cri-o://3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452" gracePeriod=2 Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.250387 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.280504 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-catalog-content\") pod \"b70a08e5-6963-4218-a193-24b52e593f7c\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.280576 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js9zs\" (UniqueName: \"kubernetes.io/projected/b70a08e5-6963-4218-a193-24b52e593f7c-kube-api-access-js9zs\") pod \"b70a08e5-6963-4218-a193-24b52e593f7c\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.280651 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-utilities\") pod \"b70a08e5-6963-4218-a193-24b52e593f7c\" (UID: \"b70a08e5-6963-4218-a193-24b52e593f7c\") " Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.281889 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-utilities" (OuterVolumeSpecName: "utilities") pod "b70a08e5-6963-4218-a193-24b52e593f7c" (UID: "b70a08e5-6963-4218-a193-24b52e593f7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.287235 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b70a08e5-6963-4218-a193-24b52e593f7c-kube-api-access-js9zs" (OuterVolumeSpecName: "kube-api-access-js9zs") pod "b70a08e5-6963-4218-a193-24b52e593f7c" (UID: "b70a08e5-6963-4218-a193-24b52e593f7c"). InnerVolumeSpecName "kube-api-access-js9zs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.324551 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b70a08e5-6963-4218-a193-24b52e593f7c" (UID: "b70a08e5-6963-4218-a193-24b52e593f7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.382562 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.382593 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js9zs\" (UniqueName: \"kubernetes.io/projected/b70a08e5-6963-4218-a193-24b52e593f7c-kube-api-access-js9zs\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.382602 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70a08e5-6963-4218-a193-24b52e593f7c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.753415 4799 generic.go:334] "Generic (PLEG): container finished" podID="b70a08e5-6963-4218-a193-24b52e593f7c" containerID="3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452" exitCode=0 Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.753502 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f2zpr" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.753498 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f2zpr" event={"ID":"b70a08e5-6963-4218-a193-24b52e593f7c","Type":"ContainerDied","Data":"3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452"} Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.753584 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f2zpr" event={"ID":"b70a08e5-6963-4218-a193-24b52e593f7c","Type":"ContainerDied","Data":"2dfb1158c952e8a47d3425e2e36f0bf5158c452a524cfd1df84fd1170eef00f2"} Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.753617 4799 scope.go:117] "RemoveContainer" containerID="3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.785490 4799 scope.go:117] "RemoveContainer" containerID="06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.804670 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f2zpr"] Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.814381 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f2zpr"] Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.819878 4799 scope.go:117] "RemoveContainer" containerID="f3460dab6b4fb4634ee2c692b9b2a8d8cd3027a02ea6bed1d7c76cc8376311be" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.890363 4799 scope.go:117] "RemoveContainer" containerID="3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452" Nov 24 08:33:36 crc kubenswrapper[4799]: E1124 08:33:36.890892 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452\": container with ID starting with 3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452 not found: ID does not exist" containerID="3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.890939 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452"} err="failed to get container status \"3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452\": rpc error: code = NotFound desc = could not find container \"3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452\": container with ID starting with 3339663e89643f7852ec984f1d6a3220cef68c3f5873faded07c191352c7f452 not found: ID does not exist" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.890995 4799 scope.go:117] "RemoveContainer" containerID="06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c" Nov 24 08:33:36 crc kubenswrapper[4799]: E1124 08:33:36.891345 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c\": container with ID starting with 06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c not found: ID does not exist" containerID="06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.891381 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c"} err="failed to get container status \"06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c\": rpc error: code = NotFound desc = could not find container \"06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c\": container with ID starting with 06db3868630b57426993ee17e5a64f1bb13e3566116c9bbda1ee05d6b679a54c not found: ID does not exist" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.891402 4799 scope.go:117] "RemoveContainer" containerID="f3460dab6b4fb4634ee2c692b9b2a8d8cd3027a02ea6bed1d7c76cc8376311be" Nov 24 08:33:36 crc kubenswrapper[4799]: E1124 08:33:36.891674 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3460dab6b4fb4634ee2c692b9b2a8d8cd3027a02ea6bed1d7c76cc8376311be\": container with ID starting with f3460dab6b4fb4634ee2c692b9b2a8d8cd3027a02ea6bed1d7c76cc8376311be not found: ID does not exist" containerID="f3460dab6b4fb4634ee2c692b9b2a8d8cd3027a02ea6bed1d7c76cc8376311be" Nov 24 08:33:36 crc kubenswrapper[4799]: I1124 08:33:36.891732 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3460dab6b4fb4634ee2c692b9b2a8d8cd3027a02ea6bed1d7c76cc8376311be"} err="failed to get container status \"f3460dab6b4fb4634ee2c692b9b2a8d8cd3027a02ea6bed1d7c76cc8376311be\": rpc error: code = NotFound desc = could not find container \"f3460dab6b4fb4634ee2c692b9b2a8d8cd3027a02ea6bed1d7c76cc8376311be\": container with ID starting with f3460dab6b4fb4634ee2c692b9b2a8d8cd3027a02ea6bed1d7c76cc8376311be not found: ID does not exist" Nov 24 08:33:37 crc kubenswrapper[4799]: I1124 08:33:37.643216 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b70a08e5-6963-4218-a193-24b52e593f7c" path="/var/lib/kubelet/pods/b70a08e5-6963-4218-a193-24b52e593f7c/volumes" Nov 24 08:33:43 crc kubenswrapper[4799]: I1124 08:33:43.628837 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:33:43 crc kubenswrapper[4799]: E1124 08:33:43.629677 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:33:55 crc kubenswrapper[4799]: I1124 08:33:55.638022 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:33:55 crc kubenswrapper[4799]: E1124 08:33:55.638967 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.082658 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wchts"] Nov 24 08:33:57 crc kubenswrapper[4799]: E1124 08:33:57.083502 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70a08e5-6963-4218-a193-24b52e593f7c" containerName="extract-content" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.083519 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70a08e5-6963-4218-a193-24b52e593f7c" containerName="extract-content" Nov 24 08:33:57 crc kubenswrapper[4799]: E1124 08:33:57.083554 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70a08e5-6963-4218-a193-24b52e593f7c" containerName="extract-utilities" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.083562 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70a08e5-6963-4218-a193-24b52e593f7c" containerName="extract-utilities" Nov 24 08:33:57 crc kubenswrapper[4799]: E1124 08:33:57.083585 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70a08e5-6963-4218-a193-24b52e593f7c" containerName="registry-server" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.083594 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70a08e5-6963-4218-a193-24b52e593f7c" containerName="registry-server" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.083966 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b70a08e5-6963-4218-a193-24b52e593f7c" containerName="registry-server" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.085974 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.112127 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wchts"] Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.225998 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-catalog-content\") pod \"redhat-operators-wchts\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.226078 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-utilities\") pod \"redhat-operators-wchts\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.226115 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5bmm\" (UniqueName: \"kubernetes.io/projected/30cbc6a5-f477-4225-a722-c92f6328b7eb-kube-api-access-j5bmm\") pod \"redhat-operators-wchts\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.328145 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-utilities\") pod \"redhat-operators-wchts\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.328232 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5bmm\" (UniqueName: \"kubernetes.io/projected/30cbc6a5-f477-4225-a722-c92f6328b7eb-kube-api-access-j5bmm\") pod \"redhat-operators-wchts\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.328421 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-catalog-content\") pod \"redhat-operators-wchts\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.328735 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-utilities\") pod \"redhat-operators-wchts\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.329002 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-catalog-content\") pod \"redhat-operators-wchts\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.372389 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5bmm\" (UniqueName: \"kubernetes.io/projected/30cbc6a5-f477-4225-a722-c92f6328b7eb-kube-api-access-j5bmm\") pod \"redhat-operators-wchts\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.410701 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.868681 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wchts"] Nov 24 08:33:57 crc kubenswrapper[4799]: I1124 08:33:57.942564 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wchts" event={"ID":"30cbc6a5-f477-4225-a722-c92f6328b7eb","Type":"ContainerStarted","Data":"7f3c5b5c74eca78288eb82f7805399d86dbb5d91f5ea6f45153d6a8c18856246"} Nov 24 08:33:58 crc kubenswrapper[4799]: I1124 08:33:58.953402 4799 generic.go:334] "Generic (PLEG): container finished" podID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerID="1c3db03a6863929ff943d0fbf1313f6df36e4e0c97a8f0c6b984a8dd759b325a" exitCode=0 Nov 24 08:33:58 crc kubenswrapper[4799]: I1124 08:33:58.953466 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wchts" event={"ID":"30cbc6a5-f477-4225-a722-c92f6328b7eb","Type":"ContainerDied","Data":"1c3db03a6863929ff943d0fbf1313f6df36e4e0c97a8f0c6b984a8dd759b325a"} Nov 24 08:33:59 crc kubenswrapper[4799]: I1124 08:33:59.965014 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wchts" event={"ID":"30cbc6a5-f477-4225-a722-c92f6328b7eb","Type":"ContainerStarted","Data":"f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326"} Nov 24 08:34:01 crc kubenswrapper[4799]: I1124 08:34:01.984741 4799 generic.go:334] "Generic (PLEG): container finished" podID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerID="f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326" exitCode=0 Nov 24 08:34:01 crc kubenswrapper[4799]: I1124 08:34:01.984790 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wchts" event={"ID":"30cbc6a5-f477-4225-a722-c92f6328b7eb","Type":"ContainerDied","Data":"f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326"} Nov 24 08:34:04 crc kubenswrapper[4799]: I1124 08:34:04.004593 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wchts" event={"ID":"30cbc6a5-f477-4225-a722-c92f6328b7eb","Type":"ContainerStarted","Data":"890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842"} Nov 24 08:34:04 crc kubenswrapper[4799]: I1124 08:34:04.026402 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wchts" podStartSLOduration=2.635577058 podStartE2EDuration="7.026385166s" podCreationTimestamp="2025-11-24 08:33:57 +0000 UTC" firstStartedPulling="2025-11-24 08:33:58.955526408 +0000 UTC m=+6384.611508892" lastFinishedPulling="2025-11-24 08:34:03.346334526 +0000 UTC m=+6389.002317000" observedRunningTime="2025-11-24 08:34:04.020252352 +0000 UTC m=+6389.676234826" watchObservedRunningTime="2025-11-24 08:34:04.026385166 +0000 UTC m=+6389.682367630" Nov 24 08:34:07 crc kubenswrapper[4799]: I1124 08:34:07.411381 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:34:07 crc kubenswrapper[4799]: I1124 08:34:07.412022 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:34:07 crc kubenswrapper[4799]: I1124 08:34:07.628378 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:34:07 crc kubenswrapper[4799]: E1124 08:34:07.628763 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:34:08 crc kubenswrapper[4799]: I1124 08:34:08.458400 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wchts" podUID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerName="registry-server" probeResult="failure" output=< Nov 24 08:34:08 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 08:34:08 crc kubenswrapper[4799]: > Nov 24 08:34:10 crc kubenswrapper[4799]: I1124 08:34:10.055405 4799 generic.go:334] "Generic (PLEG): container finished" podID="fefc2996-4fc3-40b2-bf36-82c82ba5039a" containerID="49d33ad0efcdb1ef4b49b2c386b5b14a39aa96031ee6daa726efeec118cef766" exitCode=0 Nov 24 08:34:10 crc kubenswrapper[4799]: I1124 08:34:10.055449 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" event={"ID":"fefc2996-4fc3-40b2-bf36-82c82ba5039a","Type":"ContainerDied","Data":"49d33ad0efcdb1ef4b49b2c386b5b14a39aa96031ee6daa726efeec118cef766"} Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.553567 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.612917 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-bootstrap-combined-ca-bundle\") pod \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.613012 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ssh-key\") pod \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.613168 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2bs9\" (UniqueName: \"kubernetes.io/projected/fefc2996-4fc3-40b2-bf36-82c82ba5039a-kube-api-access-t2bs9\") pod \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.613228 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-inventory\") pod \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.613334 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ceph\") pod \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\" (UID: \"fefc2996-4fc3-40b2-bf36-82c82ba5039a\") " Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.619876 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "fefc2996-4fc3-40b2-bf36-82c82ba5039a" (UID: "fefc2996-4fc3-40b2-bf36-82c82ba5039a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.620099 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fefc2996-4fc3-40b2-bf36-82c82ba5039a-kube-api-access-t2bs9" (OuterVolumeSpecName: "kube-api-access-t2bs9") pod "fefc2996-4fc3-40b2-bf36-82c82ba5039a" (UID: "fefc2996-4fc3-40b2-bf36-82c82ba5039a"). InnerVolumeSpecName "kube-api-access-t2bs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.620790 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ceph" (OuterVolumeSpecName: "ceph") pod "fefc2996-4fc3-40b2-bf36-82c82ba5039a" (UID: "fefc2996-4fc3-40b2-bf36-82c82ba5039a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.647133 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fefc2996-4fc3-40b2-bf36-82c82ba5039a" (UID: "fefc2996-4fc3-40b2-bf36-82c82ba5039a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.647651 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-inventory" (OuterVolumeSpecName: "inventory") pod "fefc2996-4fc3-40b2-bf36-82c82ba5039a" (UID: "fefc2996-4fc3-40b2-bf36-82c82ba5039a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.716856 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.716903 4799 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.716913 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.716924 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2bs9\" (UniqueName: \"kubernetes.io/projected/fefc2996-4fc3-40b2-bf36-82c82ba5039a-kube-api-access-t2bs9\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:11 crc kubenswrapper[4799]: I1124 08:34:11.716934 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fefc2996-4fc3-40b2-bf36-82c82ba5039a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.079489 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" event={"ID":"fefc2996-4fc3-40b2-bf36-82c82ba5039a","Type":"ContainerDied","Data":"8b75a48bd7e9e3b8f7aa8f22f1ed8af702d6f259e1d307ac955eaa704984fd9e"} Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.079756 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b75a48bd7e9e3b8f7aa8f22f1ed8af702d6f259e1d307ac955eaa704984fd9e" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.079547 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-lkqh5" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.172796 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-blvrs"] Nov 24 08:34:12 crc kubenswrapper[4799]: E1124 08:34:12.173369 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fefc2996-4fc3-40b2-bf36-82c82ba5039a" containerName="bootstrap-openstack-openstack-cell1" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.173391 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fefc2996-4fc3-40b2-bf36-82c82ba5039a" containerName="bootstrap-openstack-openstack-cell1" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.173608 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fefc2996-4fc3-40b2-bf36-82c82ba5039a" containerName="bootstrap-openstack-openstack-cell1" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.174336 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.176894 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.176918 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.177171 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.177517 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.193682 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-blvrs"] Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.224364 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5clp\" (UniqueName: \"kubernetes.io/projected/77211556-d2b3-4243-8518-2bfbe5ef543a-kube-api-access-v5clp\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.224518 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ceph\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.224544 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ssh-key\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.224620 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-inventory\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.326149 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5clp\" (UniqueName: \"kubernetes.io/projected/77211556-d2b3-4243-8518-2bfbe5ef543a-kube-api-access-v5clp\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.326274 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ceph\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.326305 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ssh-key\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.326402 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-inventory\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.330509 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ceph\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.330757 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ssh-key\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.342333 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-inventory\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.346676 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5clp\" (UniqueName: \"kubernetes.io/projected/77211556-d2b3-4243-8518-2bfbe5ef543a-kube-api-access-v5clp\") pod \"download-cache-openstack-openstack-cell1-blvrs\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:12 crc kubenswrapper[4799]: I1124 08:34:12.500807 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:34:13 crc kubenswrapper[4799]: I1124 08:34:13.019616 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-blvrs"] Nov 24 08:34:13 crc kubenswrapper[4799]: I1124 08:34:13.089249 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-blvrs" event={"ID":"77211556-d2b3-4243-8518-2bfbe5ef543a","Type":"ContainerStarted","Data":"28622c233dc87a0d6700e98674fd9547511c572c69f160829fdb381ac9d81f69"} Nov 24 08:34:14 crc kubenswrapper[4799]: I1124 08:34:14.099159 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-blvrs" event={"ID":"77211556-d2b3-4243-8518-2bfbe5ef543a","Type":"ContainerStarted","Data":"ceaa72234f41296cd9e3a4c9f6f35439cd1a6258169481566b80f06ac23fdd06"} Nov 24 08:34:17 crc kubenswrapper[4799]: I1124 08:34:17.460276 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:34:17 crc kubenswrapper[4799]: I1124 08:34:17.492737 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-blvrs" podStartSLOduration=4.740690447 podStartE2EDuration="5.492720821s" podCreationTimestamp="2025-11-24 08:34:12 +0000 UTC" firstStartedPulling="2025-11-24 08:34:13.022578859 +0000 UTC m=+6398.678561333" lastFinishedPulling="2025-11-24 08:34:13.774609233 +0000 UTC m=+6399.430591707" observedRunningTime="2025-11-24 08:34:14.118099782 +0000 UTC m=+6399.774082276" watchObservedRunningTime="2025-11-24 08:34:17.492720821 +0000 UTC m=+6403.148703295" Nov 24 08:34:17 crc kubenswrapper[4799]: I1124 08:34:17.511961 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:34:17 crc kubenswrapper[4799]: I1124 08:34:17.698490 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wchts"] Nov 24 08:34:19 crc kubenswrapper[4799]: I1124 08:34:19.158443 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wchts" podUID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerName="registry-server" containerID="cri-o://890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842" gracePeriod=2 Nov 24 08:34:19 crc kubenswrapper[4799]: I1124 08:34:19.654623 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:34:19 crc kubenswrapper[4799]: I1124 08:34:19.771975 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5bmm\" (UniqueName: \"kubernetes.io/projected/30cbc6a5-f477-4225-a722-c92f6328b7eb-kube-api-access-j5bmm\") pod \"30cbc6a5-f477-4225-a722-c92f6328b7eb\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " Nov 24 08:34:19 crc kubenswrapper[4799]: I1124 08:34:19.772101 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-catalog-content\") pod \"30cbc6a5-f477-4225-a722-c92f6328b7eb\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " Nov 24 08:34:19 crc kubenswrapper[4799]: I1124 08:34:19.772256 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-utilities\") pod \"30cbc6a5-f477-4225-a722-c92f6328b7eb\" (UID: \"30cbc6a5-f477-4225-a722-c92f6328b7eb\") " Nov 24 08:34:19 crc kubenswrapper[4799]: I1124 08:34:19.774691 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-utilities" (OuterVolumeSpecName: "utilities") pod "30cbc6a5-f477-4225-a722-c92f6328b7eb" (UID: "30cbc6a5-f477-4225-a722-c92f6328b7eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:34:19 crc kubenswrapper[4799]: I1124 08:34:19.778332 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30cbc6a5-f477-4225-a722-c92f6328b7eb-kube-api-access-j5bmm" (OuterVolumeSpecName: "kube-api-access-j5bmm") pod "30cbc6a5-f477-4225-a722-c92f6328b7eb" (UID: "30cbc6a5-f477-4225-a722-c92f6328b7eb"). InnerVolumeSpecName "kube-api-access-j5bmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:34:19 crc kubenswrapper[4799]: I1124 08:34:19.869862 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30cbc6a5-f477-4225-a722-c92f6328b7eb" (UID: "30cbc6a5-f477-4225-a722-c92f6328b7eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:34:19 crc kubenswrapper[4799]: I1124 08:34:19.874774 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:19 crc kubenswrapper[4799]: I1124 08:34:19.874817 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5bmm\" (UniqueName: \"kubernetes.io/projected/30cbc6a5-f477-4225-a722-c92f6328b7eb-kube-api-access-j5bmm\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:19 crc kubenswrapper[4799]: I1124 08:34:19.874830 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30cbc6a5-f477-4225-a722-c92f6328b7eb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.168375 4799 generic.go:334] "Generic (PLEG): container finished" podID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerID="890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842" exitCode=0 Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.168424 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wchts" event={"ID":"30cbc6a5-f477-4225-a722-c92f6328b7eb","Type":"ContainerDied","Data":"890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842"} Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.168649 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wchts" event={"ID":"30cbc6a5-f477-4225-a722-c92f6328b7eb","Type":"ContainerDied","Data":"7f3c5b5c74eca78288eb82f7805399d86dbb5d91f5ea6f45153d6a8c18856246"} Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.168651 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wchts" Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.168669 4799 scope.go:117] "RemoveContainer" containerID="890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842" Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.199296 4799 scope.go:117] "RemoveContainer" containerID="f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326" Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.202388 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wchts"] Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.210144 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wchts"] Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.219349 4799 scope.go:117] "RemoveContainer" containerID="1c3db03a6863929ff943d0fbf1313f6df36e4e0c97a8f0c6b984a8dd759b325a" Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.268879 4799 scope.go:117] "RemoveContainer" containerID="890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842" Nov 24 08:34:20 crc kubenswrapper[4799]: E1124 08:34:20.269621 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842\": container with ID starting with 890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842 not found: ID does not exist" containerID="890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842" Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.269665 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842"} err="failed to get container status \"890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842\": rpc error: code = NotFound desc = could not find container \"890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842\": container with ID starting with 890be17bc6f193001b02436dc25a136950d1389bd89eee7268fcaaeca83ec842 not found: ID does not exist" Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.269689 4799 scope.go:117] "RemoveContainer" containerID="f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326" Nov 24 08:34:20 crc kubenswrapper[4799]: E1124 08:34:20.269951 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326\": container with ID starting with f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326 not found: ID does not exist" containerID="f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326" Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.269984 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326"} err="failed to get container status \"f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326\": rpc error: code = NotFound desc = could not find container \"f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326\": container with ID starting with f2180ab0853946f7a03754e59e02d0e4178a588289fe7dc94779de9fcf514326 not found: ID does not exist" Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.270005 4799 scope.go:117] "RemoveContainer" containerID="1c3db03a6863929ff943d0fbf1313f6df36e4e0c97a8f0c6b984a8dd759b325a" Nov 24 08:34:20 crc kubenswrapper[4799]: E1124 08:34:20.270234 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c3db03a6863929ff943d0fbf1313f6df36e4e0c97a8f0c6b984a8dd759b325a\": container with ID starting with 1c3db03a6863929ff943d0fbf1313f6df36e4e0c97a8f0c6b984a8dd759b325a not found: ID does not exist" containerID="1c3db03a6863929ff943d0fbf1313f6df36e4e0c97a8f0c6b984a8dd759b325a" Nov 24 08:34:20 crc kubenswrapper[4799]: I1124 08:34:20.270286 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c3db03a6863929ff943d0fbf1313f6df36e4e0c97a8f0c6b984a8dd759b325a"} err="failed to get container status \"1c3db03a6863929ff943d0fbf1313f6df36e4e0c97a8f0c6b984a8dd759b325a\": rpc error: code = NotFound desc = could not find container \"1c3db03a6863929ff943d0fbf1313f6df36e4e0c97a8f0c6b984a8dd759b325a\": container with ID starting with 1c3db03a6863929ff943d0fbf1313f6df36e4e0c97a8f0c6b984a8dd759b325a not found: ID does not exist" Nov 24 08:34:21 crc kubenswrapper[4799]: I1124 08:34:21.628993 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:34:21 crc kubenswrapper[4799]: E1124 08:34:21.629349 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:34:21 crc kubenswrapper[4799]: I1124 08:34:21.641080 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30cbc6a5-f477-4225-a722-c92f6328b7eb" path="/var/lib/kubelet/pods/30cbc6a5-f477-4225-a722-c92f6328b7eb/volumes" Nov 24 08:34:33 crc kubenswrapper[4799]: I1124 08:34:33.628982 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:34:33 crc kubenswrapper[4799]: E1124 08:34:33.631288 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:34:44 crc kubenswrapper[4799]: I1124 08:34:44.629298 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:34:44 crc kubenswrapper[4799]: E1124 08:34:44.630093 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:34:58 crc kubenswrapper[4799]: I1124 08:34:58.629452 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:34:58 crc kubenswrapper[4799]: E1124 08:34:58.630879 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:35:11 crc kubenswrapper[4799]: I1124 08:35:11.628143 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:35:11 crc kubenswrapper[4799]: E1124 08:35:11.629059 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:35:25 crc kubenswrapper[4799]: I1124 08:35:25.639574 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:35:25 crc kubenswrapper[4799]: E1124 08:35:25.640960 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:35:37 crc kubenswrapper[4799]: I1124 08:35:37.628201 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:35:37 crc kubenswrapper[4799]: E1124 08:35:37.629005 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:35:46 crc kubenswrapper[4799]: I1124 08:35:46.204340 4799 generic.go:334] "Generic (PLEG): container finished" podID="77211556-d2b3-4243-8518-2bfbe5ef543a" containerID="ceaa72234f41296cd9e3a4c9f6f35439cd1a6258169481566b80f06ac23fdd06" exitCode=0 Nov 24 08:35:46 crc kubenswrapper[4799]: I1124 08:35:46.204435 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-blvrs" event={"ID":"77211556-d2b3-4243-8518-2bfbe5ef543a","Type":"ContainerDied","Data":"ceaa72234f41296cd9e3a4c9f6f35439cd1a6258169481566b80f06ac23fdd06"} Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.695677 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.813602 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5clp\" (UniqueName: \"kubernetes.io/projected/77211556-d2b3-4243-8518-2bfbe5ef543a-kube-api-access-v5clp\") pod \"77211556-d2b3-4243-8518-2bfbe5ef543a\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.813732 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ceph\") pod \"77211556-d2b3-4243-8518-2bfbe5ef543a\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.813788 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-inventory\") pod \"77211556-d2b3-4243-8518-2bfbe5ef543a\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.813805 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ssh-key\") pod \"77211556-d2b3-4243-8518-2bfbe5ef543a\" (UID: \"77211556-d2b3-4243-8518-2bfbe5ef543a\") " Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.819411 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ceph" (OuterVolumeSpecName: "ceph") pod "77211556-d2b3-4243-8518-2bfbe5ef543a" (UID: "77211556-d2b3-4243-8518-2bfbe5ef543a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.819456 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77211556-d2b3-4243-8518-2bfbe5ef543a-kube-api-access-v5clp" (OuterVolumeSpecName: "kube-api-access-v5clp") pod "77211556-d2b3-4243-8518-2bfbe5ef543a" (UID: "77211556-d2b3-4243-8518-2bfbe5ef543a"). InnerVolumeSpecName "kube-api-access-v5clp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.841801 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "77211556-d2b3-4243-8518-2bfbe5ef543a" (UID: "77211556-d2b3-4243-8518-2bfbe5ef543a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.852714 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-inventory" (OuterVolumeSpecName: "inventory") pod "77211556-d2b3-4243-8518-2bfbe5ef543a" (UID: "77211556-d2b3-4243-8518-2bfbe5ef543a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.916247 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5clp\" (UniqueName: \"kubernetes.io/projected/77211556-d2b3-4243-8518-2bfbe5ef543a-kube-api-access-v5clp\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.917144 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.917170 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:47 crc kubenswrapper[4799]: I1124 08:35:47.917185 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77211556-d2b3-4243-8518-2bfbe5ef543a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.232019 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-blvrs" event={"ID":"77211556-d2b3-4243-8518-2bfbe5ef543a","Type":"ContainerDied","Data":"28622c233dc87a0d6700e98674fd9547511c572c69f160829fdb381ac9d81f69"} Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.232087 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28622c233dc87a0d6700e98674fd9547511c572c69f160829fdb381ac9d81f69" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.232089 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-blvrs" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.330042 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-99xmg"] Nov 24 08:35:48 crc kubenswrapper[4799]: E1124 08:35:48.330427 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerName="extract-content" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.330441 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerName="extract-content" Nov 24 08:35:48 crc kubenswrapper[4799]: E1124 08:35:48.330458 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerName="extract-utilities" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.330471 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerName="extract-utilities" Nov 24 08:35:48 crc kubenswrapper[4799]: E1124 08:35:48.330495 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77211556-d2b3-4243-8518-2bfbe5ef543a" containerName="download-cache-openstack-openstack-cell1" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.330505 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="77211556-d2b3-4243-8518-2bfbe5ef543a" containerName="download-cache-openstack-openstack-cell1" Nov 24 08:35:48 crc kubenswrapper[4799]: E1124 08:35:48.330521 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerName="registry-server" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.330527 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerName="registry-server" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.330797 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="77211556-d2b3-4243-8518-2bfbe5ef543a" containerName="download-cache-openstack-openstack-cell1" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.330828 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="30cbc6a5-f477-4225-a722-c92f6328b7eb" containerName="registry-server" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.331900 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.334443 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.334747 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.335008 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.342504 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.348683 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-99xmg"] Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.425526 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn7cn\" (UniqueName: \"kubernetes.io/projected/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-kube-api-access-tn7cn\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.425631 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ssh-key\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.425835 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ceph\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.425897 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-inventory\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.527584 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ceph\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.527979 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-inventory\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.528076 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn7cn\" (UniqueName: \"kubernetes.io/projected/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-kube-api-access-tn7cn\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.528120 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ssh-key\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.532446 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ssh-key\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.532611 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-inventory\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.542962 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ceph\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.551479 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn7cn\" (UniqueName: \"kubernetes.io/projected/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-kube-api-access-tn7cn\") pod \"configure-network-openstack-openstack-cell1-99xmg\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.628259 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:35:48 crc kubenswrapper[4799]: E1124 08:35:48.628683 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:35:48 crc kubenswrapper[4799]: I1124 08:35:48.650334 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:35:49 crc kubenswrapper[4799]: I1124 08:35:49.200315 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-99xmg"] Nov 24 08:35:49 crc kubenswrapper[4799]: I1124 08:35:49.248550 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-99xmg" event={"ID":"d3b5e760-ee75-4fe6-8f12-f8cde96523c9","Type":"ContainerStarted","Data":"d71669968a4b075f95871a49f0e3f2f80b50efb28a25e46accd0bbdabb99dbce"} Nov 24 08:35:50 crc kubenswrapper[4799]: I1124 08:35:50.263302 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-99xmg" event={"ID":"d3b5e760-ee75-4fe6-8f12-f8cde96523c9","Type":"ContainerStarted","Data":"b04a8a226637c060da1fc5e9c1e40cc0033ad39912b29181a770a8801296a310"} Nov 24 08:35:50 crc kubenswrapper[4799]: I1124 08:35:50.285430 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-99xmg" podStartSLOduration=1.786449947 podStartE2EDuration="2.285403651s" podCreationTimestamp="2025-11-24 08:35:48 +0000 UTC" firstStartedPulling="2025-11-24 08:35:49.207491279 +0000 UTC m=+6494.863473753" lastFinishedPulling="2025-11-24 08:35:49.706444973 +0000 UTC m=+6495.362427457" observedRunningTime="2025-11-24 08:35:50.279452402 +0000 UTC m=+6495.935434896" watchObservedRunningTime="2025-11-24 08:35:50.285403651 +0000 UTC m=+6495.941386125" Nov 24 08:36:02 crc kubenswrapper[4799]: I1124 08:36:02.629579 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:36:02 crc kubenswrapper[4799]: E1124 08:36:02.630403 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:36:15 crc kubenswrapper[4799]: I1124 08:36:15.634796 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:36:15 crc kubenswrapper[4799]: E1124 08:36:15.636548 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:36:27 crc kubenswrapper[4799]: I1124 08:36:27.629201 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:36:27 crc kubenswrapper[4799]: E1124 08:36:27.630265 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.368670 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hh7sl"] Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.373830 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.381238 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hh7sl"] Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.520280 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swthc\" (UniqueName: \"kubernetes.io/projected/f7d8ed9c-3246-489a-9c5f-ea9972265038-kube-api-access-swthc\") pod \"redhat-marketplace-hh7sl\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.520419 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-catalog-content\") pod \"redhat-marketplace-hh7sl\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.520454 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-utilities\") pod \"redhat-marketplace-hh7sl\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.623091 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swthc\" (UniqueName: \"kubernetes.io/projected/f7d8ed9c-3246-489a-9c5f-ea9972265038-kube-api-access-swthc\") pod \"redhat-marketplace-hh7sl\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.624659 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-catalog-content\") pod \"redhat-marketplace-hh7sl\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.625005 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-utilities\") pod \"redhat-marketplace-hh7sl\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.625811 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-utilities\") pod \"redhat-marketplace-hh7sl\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.626135 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-catalog-content\") pod \"redhat-marketplace-hh7sl\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.643327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swthc\" (UniqueName: \"kubernetes.io/projected/f7d8ed9c-3246-489a-9c5f-ea9972265038-kube-api-access-swthc\") pod \"redhat-marketplace-hh7sl\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:35 crc kubenswrapper[4799]: I1124 08:36:35.704457 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:36 crc kubenswrapper[4799]: I1124 08:36:36.350781 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hh7sl"] Nov 24 08:36:36 crc kubenswrapper[4799]: I1124 08:36:36.765623 4799 generic.go:334] "Generic (PLEG): container finished" podID="f7d8ed9c-3246-489a-9c5f-ea9972265038" containerID="eda99aebead594dfc9574e740411e2d5a6e5155965ce96e03713bf53c4abcbca" exitCode=0 Nov 24 08:36:36 crc kubenswrapper[4799]: I1124 08:36:36.765666 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hh7sl" event={"ID":"f7d8ed9c-3246-489a-9c5f-ea9972265038","Type":"ContainerDied","Data":"eda99aebead594dfc9574e740411e2d5a6e5155965ce96e03713bf53c4abcbca"} Nov 24 08:36:36 crc kubenswrapper[4799]: I1124 08:36:36.765704 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hh7sl" event={"ID":"f7d8ed9c-3246-489a-9c5f-ea9972265038","Type":"ContainerStarted","Data":"58b00e70bd7115edd4439e5837c353d95ecd8836f48505425c4ba60a0c481e0f"} Nov 24 08:36:36 crc kubenswrapper[4799]: I1124 08:36:36.768007 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:36:37 crc kubenswrapper[4799]: I1124 08:36:37.777328 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hh7sl" event={"ID":"f7d8ed9c-3246-489a-9c5f-ea9972265038","Type":"ContainerStarted","Data":"2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5"} Nov 24 08:36:38 crc kubenswrapper[4799]: I1124 08:36:38.791697 4799 generic.go:334] "Generic (PLEG): container finished" podID="f7d8ed9c-3246-489a-9c5f-ea9972265038" containerID="2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5" exitCode=0 Nov 24 08:36:38 crc kubenswrapper[4799]: I1124 08:36:38.791754 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hh7sl" event={"ID":"f7d8ed9c-3246-489a-9c5f-ea9972265038","Type":"ContainerDied","Data":"2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5"} Nov 24 08:36:39 crc kubenswrapper[4799]: I1124 08:36:39.806743 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hh7sl" event={"ID":"f7d8ed9c-3246-489a-9c5f-ea9972265038","Type":"ContainerStarted","Data":"3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e"} Nov 24 08:36:39 crc kubenswrapper[4799]: I1124 08:36:39.832129 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hh7sl" podStartSLOduration=2.387562065 podStartE2EDuration="4.832106852s" podCreationTimestamp="2025-11-24 08:36:35 +0000 UTC" firstStartedPulling="2025-11-24 08:36:36.767742148 +0000 UTC m=+6542.423724622" lastFinishedPulling="2025-11-24 08:36:39.212286915 +0000 UTC m=+6544.868269409" observedRunningTime="2025-11-24 08:36:39.824836626 +0000 UTC m=+6545.480819110" watchObservedRunningTime="2025-11-24 08:36:39.832106852 +0000 UTC m=+6545.488089326" Nov 24 08:36:41 crc kubenswrapper[4799]: I1124 08:36:41.629801 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:36:41 crc kubenswrapper[4799]: E1124 08:36:41.630964 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:36:45 crc kubenswrapper[4799]: I1124 08:36:45.705680 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:45 crc kubenswrapper[4799]: I1124 08:36:45.707133 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:45 crc kubenswrapper[4799]: I1124 08:36:45.752607 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:45 crc kubenswrapper[4799]: I1124 08:36:45.918543 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:45 crc kubenswrapper[4799]: I1124 08:36:45.985028 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hh7sl"] Nov 24 08:36:47 crc kubenswrapper[4799]: I1124 08:36:47.884663 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hh7sl" podUID="f7d8ed9c-3246-489a-9c5f-ea9972265038" containerName="registry-server" containerID="cri-o://3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e" gracePeriod=2 Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.388224 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.507371 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-catalog-content\") pod \"f7d8ed9c-3246-489a-9c5f-ea9972265038\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.507470 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-utilities\") pod \"f7d8ed9c-3246-489a-9c5f-ea9972265038\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.507641 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swthc\" (UniqueName: \"kubernetes.io/projected/f7d8ed9c-3246-489a-9c5f-ea9972265038-kube-api-access-swthc\") pod \"f7d8ed9c-3246-489a-9c5f-ea9972265038\" (UID: \"f7d8ed9c-3246-489a-9c5f-ea9972265038\") " Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.508633 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-utilities" (OuterVolumeSpecName: "utilities") pod "f7d8ed9c-3246-489a-9c5f-ea9972265038" (UID: "f7d8ed9c-3246-489a-9c5f-ea9972265038"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.516271 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7d8ed9c-3246-489a-9c5f-ea9972265038-kube-api-access-swthc" (OuterVolumeSpecName: "kube-api-access-swthc") pod "f7d8ed9c-3246-489a-9c5f-ea9972265038" (UID: "f7d8ed9c-3246-489a-9c5f-ea9972265038"). InnerVolumeSpecName "kube-api-access-swthc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.525481 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7d8ed9c-3246-489a-9c5f-ea9972265038" (UID: "f7d8ed9c-3246-489a-9c5f-ea9972265038"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.609567 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swthc\" (UniqueName: \"kubernetes.io/projected/f7d8ed9c-3246-489a-9c5f-ea9972265038-kube-api-access-swthc\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.609596 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.609608 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7d8ed9c-3246-489a-9c5f-ea9972265038-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.897584 4799 generic.go:334] "Generic (PLEG): container finished" podID="f7d8ed9c-3246-489a-9c5f-ea9972265038" containerID="3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e" exitCode=0 Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.897639 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hh7sl" event={"ID":"f7d8ed9c-3246-489a-9c5f-ea9972265038","Type":"ContainerDied","Data":"3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e"} Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.897689 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hh7sl" Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.897717 4799 scope.go:117] "RemoveContainer" containerID="3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e" Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.897704 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hh7sl" event={"ID":"f7d8ed9c-3246-489a-9c5f-ea9972265038","Type":"ContainerDied","Data":"58b00e70bd7115edd4439e5837c353d95ecd8836f48505425c4ba60a0c481e0f"} Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.928372 4799 scope.go:117] "RemoveContainer" containerID="2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5" Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.944156 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hh7sl"] Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.955564 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hh7sl"] Nov 24 08:36:48 crc kubenswrapper[4799]: I1124 08:36:48.959999 4799 scope.go:117] "RemoveContainer" containerID="eda99aebead594dfc9574e740411e2d5a6e5155965ce96e03713bf53c4abcbca" Nov 24 08:36:49 crc kubenswrapper[4799]: I1124 08:36:49.004916 4799 scope.go:117] "RemoveContainer" containerID="3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e" Nov 24 08:36:49 crc kubenswrapper[4799]: E1124 08:36:49.005728 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e\": container with ID starting with 3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e not found: ID does not exist" containerID="3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e" Nov 24 08:36:49 crc kubenswrapper[4799]: I1124 08:36:49.005810 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e"} err="failed to get container status \"3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e\": rpc error: code = NotFound desc = could not find container \"3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e\": container with ID starting with 3be5988219e7e752c63d7ab70bf43d9d9067bdc4bbf6fe480249442fc2af872e not found: ID does not exist" Nov 24 08:36:49 crc kubenswrapper[4799]: I1124 08:36:49.005839 4799 scope.go:117] "RemoveContainer" containerID="2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5" Nov 24 08:36:49 crc kubenswrapper[4799]: E1124 08:36:49.006569 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5\": container with ID starting with 2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5 not found: ID does not exist" containerID="2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5" Nov 24 08:36:49 crc kubenswrapper[4799]: I1124 08:36:49.006634 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5"} err="failed to get container status \"2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5\": rpc error: code = NotFound desc = could not find container \"2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5\": container with ID starting with 2369ee05d52379fbc0b86ef541840fc5a023fa5dd4629c334347e2fea3aea3a5 not found: ID does not exist" Nov 24 08:36:49 crc kubenswrapper[4799]: I1124 08:36:49.006678 4799 scope.go:117] "RemoveContainer" containerID="eda99aebead594dfc9574e740411e2d5a6e5155965ce96e03713bf53c4abcbca" Nov 24 08:36:49 crc kubenswrapper[4799]: E1124 08:36:49.007188 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eda99aebead594dfc9574e740411e2d5a6e5155965ce96e03713bf53c4abcbca\": container with ID starting with eda99aebead594dfc9574e740411e2d5a6e5155965ce96e03713bf53c4abcbca not found: ID does not exist" containerID="eda99aebead594dfc9574e740411e2d5a6e5155965ce96e03713bf53c4abcbca" Nov 24 08:36:49 crc kubenswrapper[4799]: I1124 08:36:49.007233 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eda99aebead594dfc9574e740411e2d5a6e5155965ce96e03713bf53c4abcbca"} err="failed to get container status \"eda99aebead594dfc9574e740411e2d5a6e5155965ce96e03713bf53c4abcbca\": rpc error: code = NotFound desc = could not find container \"eda99aebead594dfc9574e740411e2d5a6e5155965ce96e03713bf53c4abcbca\": container with ID starting with eda99aebead594dfc9574e740411e2d5a6e5155965ce96e03713bf53c4abcbca not found: ID does not exist" Nov 24 08:36:49 crc kubenswrapper[4799]: I1124 08:36:49.643092 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7d8ed9c-3246-489a-9c5f-ea9972265038" path="/var/lib/kubelet/pods/f7d8ed9c-3246-489a-9c5f-ea9972265038/volumes" Nov 24 08:36:54 crc kubenswrapper[4799]: I1124 08:36:54.628818 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:36:54 crc kubenswrapper[4799]: E1124 08:36:54.629996 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:37:08 crc kubenswrapper[4799]: I1124 08:37:08.628974 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:37:08 crc kubenswrapper[4799]: E1124 08:37:08.630838 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:37:12 crc kubenswrapper[4799]: I1124 08:37:12.126565 4799 generic.go:334] "Generic (PLEG): container finished" podID="d3b5e760-ee75-4fe6-8f12-f8cde96523c9" containerID="b04a8a226637c060da1fc5e9c1e40cc0033ad39912b29181a770a8801296a310" exitCode=0 Nov 24 08:37:12 crc kubenswrapper[4799]: I1124 08:37:12.126643 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-99xmg" event={"ID":"d3b5e760-ee75-4fe6-8f12-f8cde96523c9","Type":"ContainerDied","Data":"b04a8a226637c060da1fc5e9c1e40cc0033ad39912b29181a770a8801296a310"} Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.607281 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.741092 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ceph\") pod \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.741211 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ssh-key\") pod \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.741312 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tn7cn\" (UniqueName: \"kubernetes.io/projected/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-kube-api-access-tn7cn\") pod \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.741482 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-inventory\") pod \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\" (UID: \"d3b5e760-ee75-4fe6-8f12-f8cde96523c9\") " Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.779148 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ceph" (OuterVolumeSpecName: "ceph") pod "d3b5e760-ee75-4fe6-8f12-f8cde96523c9" (UID: "d3b5e760-ee75-4fe6-8f12-f8cde96523c9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.791322 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-kube-api-access-tn7cn" (OuterVolumeSpecName: "kube-api-access-tn7cn") pod "d3b5e760-ee75-4fe6-8f12-f8cde96523c9" (UID: "d3b5e760-ee75-4fe6-8f12-f8cde96523c9"). InnerVolumeSpecName "kube-api-access-tn7cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.813021 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-inventory" (OuterVolumeSpecName: "inventory") pod "d3b5e760-ee75-4fe6-8f12-f8cde96523c9" (UID: "d3b5e760-ee75-4fe6-8f12-f8cde96523c9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.825053 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d3b5e760-ee75-4fe6-8f12-f8cde96523c9" (UID: "d3b5e760-ee75-4fe6-8f12-f8cde96523c9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.847159 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.847191 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.847201 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:13 crc kubenswrapper[4799]: I1124 08:37:13.847211 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tn7cn\" (UniqueName: \"kubernetes.io/projected/d3b5e760-ee75-4fe6-8f12-f8cde96523c9-kube-api-access-tn7cn\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.152430 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-99xmg" event={"ID":"d3b5e760-ee75-4fe6-8f12-f8cde96523c9","Type":"ContainerDied","Data":"d71669968a4b075f95871a49f0e3f2f80b50efb28a25e46accd0bbdabb99dbce"} Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.152477 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d71669968a4b075f95871a49f0e3f2f80b50efb28a25e46accd0bbdabb99dbce" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.152547 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-99xmg" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.240163 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-9hl97"] Nov 24 08:37:14 crc kubenswrapper[4799]: E1124 08:37:14.240616 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d8ed9c-3246-489a-9c5f-ea9972265038" containerName="extract-utilities" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.240633 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d8ed9c-3246-489a-9c5f-ea9972265038" containerName="extract-utilities" Nov 24 08:37:14 crc kubenswrapper[4799]: E1124 08:37:14.240645 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d8ed9c-3246-489a-9c5f-ea9972265038" containerName="extract-content" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.240651 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d8ed9c-3246-489a-9c5f-ea9972265038" containerName="extract-content" Nov 24 08:37:14 crc kubenswrapper[4799]: E1124 08:37:14.240690 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3b5e760-ee75-4fe6-8f12-f8cde96523c9" containerName="configure-network-openstack-openstack-cell1" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.240697 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3b5e760-ee75-4fe6-8f12-f8cde96523c9" containerName="configure-network-openstack-openstack-cell1" Nov 24 08:37:14 crc kubenswrapper[4799]: E1124 08:37:14.240707 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d8ed9c-3246-489a-9c5f-ea9972265038" containerName="registry-server" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.240713 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d8ed9c-3246-489a-9c5f-ea9972265038" containerName="registry-server" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.240927 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3b5e760-ee75-4fe6-8f12-f8cde96523c9" containerName="configure-network-openstack-openstack-cell1" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.240952 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7d8ed9c-3246-489a-9c5f-ea9972265038" containerName="registry-server" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.241772 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.245469 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.245590 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.249938 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.249961 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.263738 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-9hl97"] Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.358497 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-inventory\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.358580 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ssh-key\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.358601 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ceph\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.358650 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7gzw\" (UniqueName: \"kubernetes.io/projected/525cb867-093d-4379-b617-0590c59f6ef1-kube-api-access-z7gzw\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.460676 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-inventory\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.461080 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ssh-key\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.461109 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ceph\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.461160 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7gzw\" (UniqueName: \"kubernetes.io/projected/525cb867-093d-4379-b617-0590c59f6ef1-kube-api-access-z7gzw\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.465066 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ssh-key\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.465142 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ceph\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.465632 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-inventory\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.477020 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7gzw\" (UniqueName: \"kubernetes.io/projected/525cb867-093d-4379-b617-0590c59f6ef1-kube-api-access-z7gzw\") pod \"validate-network-openstack-openstack-cell1-9hl97\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:14 crc kubenswrapper[4799]: I1124 08:37:14.559486 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:15 crc kubenswrapper[4799]: I1124 08:37:15.086764 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-9hl97"] Nov 24 08:37:15 crc kubenswrapper[4799]: I1124 08:37:15.164243 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-9hl97" event={"ID":"525cb867-093d-4379-b617-0590c59f6ef1","Type":"ContainerStarted","Data":"4878e64dd07e6ab3011269d6d2b342ebcf76a3bf64d88e8dfc3bf41a3969cd3c"} Nov 24 08:37:16 crc kubenswrapper[4799]: I1124 08:37:16.174154 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-9hl97" event={"ID":"525cb867-093d-4379-b617-0590c59f6ef1","Type":"ContainerStarted","Data":"1f44d5bc79383998670ad0b92e0f76b7ed7de56af9efe5721318620df4f1fa10"} Nov 24 08:37:16 crc kubenswrapper[4799]: I1124 08:37:16.197600 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-9hl97" podStartSLOduration=1.714665777 podStartE2EDuration="2.197583215s" podCreationTimestamp="2025-11-24 08:37:14 +0000 UTC" firstStartedPulling="2025-11-24 08:37:15.095052175 +0000 UTC m=+6580.751034649" lastFinishedPulling="2025-11-24 08:37:15.577969573 +0000 UTC m=+6581.233952087" observedRunningTime="2025-11-24 08:37:16.190964637 +0000 UTC m=+6581.846947121" watchObservedRunningTime="2025-11-24 08:37:16.197583215 +0000 UTC m=+6581.853565689" Nov 24 08:37:21 crc kubenswrapper[4799]: I1124 08:37:21.220899 4799 generic.go:334] "Generic (PLEG): container finished" podID="525cb867-093d-4379-b617-0590c59f6ef1" containerID="1f44d5bc79383998670ad0b92e0f76b7ed7de56af9efe5721318620df4f1fa10" exitCode=0 Nov 24 08:37:21 crc kubenswrapper[4799]: I1124 08:37:21.221559 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-9hl97" event={"ID":"525cb867-093d-4379-b617-0590c59f6ef1","Type":"ContainerDied","Data":"1f44d5bc79383998670ad0b92e0f76b7ed7de56af9efe5721318620df4f1fa10"} Nov 24 08:37:21 crc kubenswrapper[4799]: I1124 08:37:21.629333 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:37:21 crc kubenswrapper[4799]: E1124 08:37:21.629874 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.704743 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.769305 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ceph\") pod \"525cb867-093d-4379-b617-0590c59f6ef1\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.769388 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ssh-key\") pod \"525cb867-093d-4379-b617-0590c59f6ef1\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.769664 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-inventory\") pod \"525cb867-093d-4379-b617-0590c59f6ef1\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.769776 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7gzw\" (UniqueName: \"kubernetes.io/projected/525cb867-093d-4379-b617-0590c59f6ef1-kube-api-access-z7gzw\") pod \"525cb867-093d-4379-b617-0590c59f6ef1\" (UID: \"525cb867-093d-4379-b617-0590c59f6ef1\") " Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.778949 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/525cb867-093d-4379-b617-0590c59f6ef1-kube-api-access-z7gzw" (OuterVolumeSpecName: "kube-api-access-z7gzw") pod "525cb867-093d-4379-b617-0590c59f6ef1" (UID: "525cb867-093d-4379-b617-0590c59f6ef1"). InnerVolumeSpecName "kube-api-access-z7gzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.784181 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ceph" (OuterVolumeSpecName: "ceph") pod "525cb867-093d-4379-b617-0590c59f6ef1" (UID: "525cb867-093d-4379-b617-0590c59f6ef1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.800367 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "525cb867-093d-4379-b617-0590c59f6ef1" (UID: "525cb867-093d-4379-b617-0590c59f6ef1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.816746 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-inventory" (OuterVolumeSpecName: "inventory") pod "525cb867-093d-4379-b617-0590c59f6ef1" (UID: "525cb867-093d-4379-b617-0590c59f6ef1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.873393 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.873421 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7gzw\" (UniqueName: \"kubernetes.io/projected/525cb867-093d-4379-b617-0590c59f6ef1-kube-api-access-z7gzw\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.873432 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:22 crc kubenswrapper[4799]: I1124 08:37:22.873441 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/525cb867-093d-4379-b617-0590c59f6ef1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.244101 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-9hl97" event={"ID":"525cb867-093d-4379-b617-0590c59f6ef1","Type":"ContainerDied","Data":"4878e64dd07e6ab3011269d6d2b342ebcf76a3bf64d88e8dfc3bf41a3969cd3c"} Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.244163 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4878e64dd07e6ab3011269d6d2b342ebcf76a3bf64d88e8dfc3bf41a3969cd3c" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.244184 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-9hl97" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.344139 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-9nqzd"] Nov 24 08:37:23 crc kubenswrapper[4799]: E1124 08:37:23.344917 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="525cb867-093d-4379-b617-0590c59f6ef1" containerName="validate-network-openstack-openstack-cell1" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.344939 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="525cb867-093d-4379-b617-0590c59f6ef1" containerName="validate-network-openstack-openstack-cell1" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.345153 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="525cb867-093d-4379-b617-0590c59f6ef1" containerName="validate-network-openstack-openstack-cell1" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.345872 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.347960 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.348109 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.348568 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.360795 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.366383 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-9nqzd"] Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.384558 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ceph\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.384658 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24glg\" (UniqueName: \"kubernetes.io/projected/b80d7525-1ef0-4824-857a-5264a31f9b73-kube-api-access-24glg\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.384900 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-inventory\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.384936 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ssh-key\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.486000 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-inventory\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.486047 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ssh-key\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.486136 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ceph\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.486177 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24glg\" (UniqueName: \"kubernetes.io/projected/b80d7525-1ef0-4824-857a-5264a31f9b73-kube-api-access-24glg\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.490707 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-inventory\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.490759 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ceph\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.491494 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ssh-key\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.508542 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24glg\" (UniqueName: \"kubernetes.io/projected/b80d7525-1ef0-4824-857a-5264a31f9b73-kube-api-access-24glg\") pod \"install-os-openstack-openstack-cell1-9nqzd\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:23 crc kubenswrapper[4799]: I1124 08:37:23.719516 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:37:24 crc kubenswrapper[4799]: I1124 08:37:24.271869 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-9nqzd"] Nov 24 08:37:25 crc kubenswrapper[4799]: I1124 08:37:25.266090 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-9nqzd" event={"ID":"b80d7525-1ef0-4824-857a-5264a31f9b73","Type":"ContainerStarted","Data":"2f8fddf2362f7dcb5974f926f5031b0288e10704156fe33b6d3a82a57c81273f"} Nov 24 08:37:25 crc kubenswrapper[4799]: I1124 08:37:25.266697 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-9nqzd" event={"ID":"b80d7525-1ef0-4824-857a-5264a31f9b73","Type":"ContainerStarted","Data":"ebd9db4d3bfd046cfc6acac91a99856ac8c1819a20d271089086748fcc85c69a"} Nov 24 08:37:25 crc kubenswrapper[4799]: I1124 08:37:25.289595 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-9nqzd" podStartSLOduration=1.87072247 podStartE2EDuration="2.289573939s" podCreationTimestamp="2025-11-24 08:37:23 +0000 UTC" firstStartedPulling="2025-11-24 08:37:24.276714705 +0000 UTC m=+6589.932697179" lastFinishedPulling="2025-11-24 08:37:24.695566174 +0000 UTC m=+6590.351548648" observedRunningTime="2025-11-24 08:37:25.281668785 +0000 UTC m=+6590.937651269" watchObservedRunningTime="2025-11-24 08:37:25.289573939 +0000 UTC m=+6590.945556413" Nov 24 08:37:34 crc kubenswrapper[4799]: I1124 08:37:34.628550 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:37:34 crc kubenswrapper[4799]: E1124 08:37:34.629445 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:37:49 crc kubenswrapper[4799]: I1124 08:37:49.629143 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:37:49 crc kubenswrapper[4799]: E1124 08:37:49.629938 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.229645 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cm6v8"] Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.238107 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.244471 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cm6v8"] Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.384213 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-catalog-content\") pod \"community-operators-cm6v8\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.384468 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f64t5\" (UniqueName: \"kubernetes.io/projected/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-kube-api-access-f64t5\") pod \"community-operators-cm6v8\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.384642 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-utilities\") pod \"community-operators-cm6v8\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.486097 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-utilities\") pod \"community-operators-cm6v8\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.486229 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-catalog-content\") pod \"community-operators-cm6v8\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.486272 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f64t5\" (UniqueName: \"kubernetes.io/projected/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-kube-api-access-f64t5\") pod \"community-operators-cm6v8\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.487087 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-utilities\") pod \"community-operators-cm6v8\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.487324 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-catalog-content\") pod \"community-operators-cm6v8\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.517895 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f64t5\" (UniqueName: \"kubernetes.io/projected/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-kube-api-access-f64t5\") pod \"community-operators-cm6v8\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:37:53 crc kubenswrapper[4799]: I1124 08:37:53.598453 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:37:54 crc kubenswrapper[4799]: I1124 08:37:54.105564 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cm6v8"] Nov 24 08:37:54 crc kubenswrapper[4799]: I1124 08:37:54.571192 4799 generic.go:334] "Generic (PLEG): container finished" podID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" containerID="30291ef2d24403106ee4896e5f98e1fc14b7e55b9923d9819739d728739bde04" exitCode=0 Nov 24 08:37:54 crc kubenswrapper[4799]: I1124 08:37:54.571250 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm6v8" event={"ID":"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343","Type":"ContainerDied","Data":"30291ef2d24403106ee4896e5f98e1fc14b7e55b9923d9819739d728739bde04"} Nov 24 08:37:54 crc kubenswrapper[4799]: I1124 08:37:54.571300 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm6v8" event={"ID":"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343","Type":"ContainerStarted","Data":"0faf2658b38e65f9305f61b7a481472e0a7f6f59d1ff1bb3af2136f7faabace6"} Nov 24 08:37:56 crc kubenswrapper[4799]: I1124 08:37:56.589939 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm6v8" event={"ID":"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343","Type":"ContainerStarted","Data":"dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792"} Nov 24 08:37:57 crc kubenswrapper[4799]: I1124 08:37:57.602288 4799 generic.go:334] "Generic (PLEG): container finished" podID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" containerID="dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792" exitCode=0 Nov 24 08:37:57 crc kubenswrapper[4799]: I1124 08:37:57.602407 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm6v8" event={"ID":"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343","Type":"ContainerDied","Data":"dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792"} Nov 24 08:37:58 crc kubenswrapper[4799]: I1124 08:37:58.616444 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm6v8" event={"ID":"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343","Type":"ContainerStarted","Data":"baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807"} Nov 24 08:37:58 crc kubenswrapper[4799]: I1124 08:37:58.652009 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cm6v8" podStartSLOduration=2.112840874 podStartE2EDuration="5.651981367s" podCreationTimestamp="2025-11-24 08:37:53 +0000 UTC" firstStartedPulling="2025-11-24 08:37:54.573298528 +0000 UTC m=+6620.229281002" lastFinishedPulling="2025-11-24 08:37:58.112439021 +0000 UTC m=+6623.768421495" observedRunningTime="2025-11-24 08:37:58.648313523 +0000 UTC m=+6624.304295997" watchObservedRunningTime="2025-11-24 08:37:58.651981367 +0000 UTC m=+6624.307963861" Nov 24 08:38:03 crc kubenswrapper[4799]: I1124 08:38:03.598944 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:38:03 crc kubenswrapper[4799]: I1124 08:38:03.599379 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:38:03 crc kubenswrapper[4799]: I1124 08:38:03.629539 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:38:03 crc kubenswrapper[4799]: E1124 08:38:03.629828 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:38:03 crc kubenswrapper[4799]: I1124 08:38:03.648956 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:38:03 crc kubenswrapper[4799]: I1124 08:38:03.735544 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:38:04 crc kubenswrapper[4799]: I1124 08:38:04.226619 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cm6v8"] Nov 24 08:38:05 crc kubenswrapper[4799]: I1124 08:38:05.703486 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cm6v8" podUID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" containerName="registry-server" containerID="cri-o://baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807" gracePeriod=2 Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.200349 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.363806 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-catalog-content\") pod \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.363924 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-utilities\") pod \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.364095 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f64t5\" (UniqueName: \"kubernetes.io/projected/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-kube-api-access-f64t5\") pod \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\" (UID: \"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343\") " Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.364911 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-utilities" (OuterVolumeSpecName: "utilities") pod "3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" (UID: "3bcb322a-ed70-4d00-b5b6-a3cfe1e93343"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.387202 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-kube-api-access-f64t5" (OuterVolumeSpecName: "kube-api-access-f64t5") pod "3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" (UID: "3bcb322a-ed70-4d00-b5b6-a3cfe1e93343"). InnerVolumeSpecName "kube-api-access-f64t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.479827 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.479917 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f64t5\" (UniqueName: \"kubernetes.io/projected/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-kube-api-access-f64t5\") on node \"crc\" DevicePath \"\"" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.725433 4799 generic.go:334] "Generic (PLEG): container finished" podID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" containerID="baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807" exitCode=0 Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.725470 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm6v8" event={"ID":"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343","Type":"ContainerDied","Data":"baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807"} Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.725516 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm6v8" event={"ID":"3bcb322a-ed70-4d00-b5b6-a3cfe1e93343","Type":"ContainerDied","Data":"0faf2658b38e65f9305f61b7a481472e0a7f6f59d1ff1bb3af2136f7faabace6"} Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.725517 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cm6v8" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.725534 4799 scope.go:117] "RemoveContainer" containerID="baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.750993 4799 scope.go:117] "RemoveContainer" containerID="dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.775190 4799 scope.go:117] "RemoveContainer" containerID="30291ef2d24403106ee4896e5f98e1fc14b7e55b9923d9819739d728739bde04" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.803997 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" (UID: "3bcb322a-ed70-4d00-b5b6-a3cfe1e93343"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.823048 4799 scope.go:117] "RemoveContainer" containerID="baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807" Nov 24 08:38:06 crc kubenswrapper[4799]: E1124 08:38:06.823703 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807\": container with ID starting with baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807 not found: ID does not exist" containerID="baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.823775 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807"} err="failed to get container status \"baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807\": rpc error: code = NotFound desc = could not find container \"baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807\": container with ID starting with baa2f636840e0e967b36647b2580643b8f032fb1c847eac00189f5ea524c7807 not found: ID does not exist" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.823815 4799 scope.go:117] "RemoveContainer" containerID="dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792" Nov 24 08:38:06 crc kubenswrapper[4799]: E1124 08:38:06.825300 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792\": container with ID starting with dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792 not found: ID does not exist" containerID="dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.825348 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792"} err="failed to get container status \"dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792\": rpc error: code = NotFound desc = could not find container \"dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792\": container with ID starting with dee5540ca156e6a8553f9d8e520a40ac6c32124da6a75879a3d0dc5ce7d9e792 not found: ID does not exist" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.825376 4799 scope.go:117] "RemoveContainer" containerID="30291ef2d24403106ee4896e5f98e1fc14b7e55b9923d9819739d728739bde04" Nov 24 08:38:06 crc kubenswrapper[4799]: E1124 08:38:06.825707 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30291ef2d24403106ee4896e5f98e1fc14b7e55b9923d9819739d728739bde04\": container with ID starting with 30291ef2d24403106ee4896e5f98e1fc14b7e55b9923d9819739d728739bde04 not found: ID does not exist" containerID="30291ef2d24403106ee4896e5f98e1fc14b7e55b9923d9819739d728739bde04" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.825732 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30291ef2d24403106ee4896e5f98e1fc14b7e55b9923d9819739d728739bde04"} err="failed to get container status \"30291ef2d24403106ee4896e5f98e1fc14b7e55b9923d9819739d728739bde04\": rpc error: code = NotFound desc = could not find container \"30291ef2d24403106ee4896e5f98e1fc14b7e55b9923d9819739d728739bde04\": container with ID starting with 30291ef2d24403106ee4896e5f98e1fc14b7e55b9923d9819739d728739bde04 not found: ID does not exist" Nov 24 08:38:06 crc kubenswrapper[4799]: I1124 08:38:06.888518 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:38:07 crc kubenswrapper[4799]: I1124 08:38:07.066651 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cm6v8"] Nov 24 08:38:07 crc kubenswrapper[4799]: I1124 08:38:07.083050 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cm6v8"] Nov 24 08:38:07 crc kubenswrapper[4799]: I1124 08:38:07.641002 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" path="/var/lib/kubelet/pods/3bcb322a-ed70-4d00-b5b6-a3cfe1e93343/volumes" Nov 24 08:38:09 crc kubenswrapper[4799]: I1124 08:38:09.759234 4799 generic.go:334] "Generic (PLEG): container finished" podID="b80d7525-1ef0-4824-857a-5264a31f9b73" containerID="2f8fddf2362f7dcb5974f926f5031b0288e10704156fe33b6d3a82a57c81273f" exitCode=0 Nov 24 08:38:09 crc kubenswrapper[4799]: I1124 08:38:09.759469 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-9nqzd" event={"ID":"b80d7525-1ef0-4824-857a-5264a31f9b73","Type":"ContainerDied","Data":"2f8fddf2362f7dcb5974f926f5031b0288e10704156fe33b6d3a82a57c81273f"} Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.248206 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.401604 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ceph\") pod \"b80d7525-1ef0-4824-857a-5264a31f9b73\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.401769 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-inventory\") pod \"b80d7525-1ef0-4824-857a-5264a31f9b73\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.401867 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24glg\" (UniqueName: \"kubernetes.io/projected/b80d7525-1ef0-4824-857a-5264a31f9b73-kube-api-access-24glg\") pod \"b80d7525-1ef0-4824-857a-5264a31f9b73\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.401964 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ssh-key\") pod \"b80d7525-1ef0-4824-857a-5264a31f9b73\" (UID: \"b80d7525-1ef0-4824-857a-5264a31f9b73\") " Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.409145 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ceph" (OuterVolumeSpecName: "ceph") pod "b80d7525-1ef0-4824-857a-5264a31f9b73" (UID: "b80d7525-1ef0-4824-857a-5264a31f9b73"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.413520 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80d7525-1ef0-4824-857a-5264a31f9b73-kube-api-access-24glg" (OuterVolumeSpecName: "kube-api-access-24glg") pod "b80d7525-1ef0-4824-857a-5264a31f9b73" (UID: "b80d7525-1ef0-4824-857a-5264a31f9b73"). InnerVolumeSpecName "kube-api-access-24glg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.435805 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-inventory" (OuterVolumeSpecName: "inventory") pod "b80d7525-1ef0-4824-857a-5264a31f9b73" (UID: "b80d7525-1ef0-4824-857a-5264a31f9b73"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.438821 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b80d7525-1ef0-4824-857a-5264a31f9b73" (UID: "b80d7525-1ef0-4824-857a-5264a31f9b73"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.504149 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24glg\" (UniqueName: \"kubernetes.io/projected/b80d7525-1ef0-4824-857a-5264a31f9b73-kube-api-access-24glg\") on node \"crc\" DevicePath \"\"" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.504205 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.504220 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.504235 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b80d7525-1ef0-4824-857a-5264a31f9b73-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.779260 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-9nqzd" event={"ID":"b80d7525-1ef0-4824-857a-5264a31f9b73","Type":"ContainerDied","Data":"ebd9db4d3bfd046cfc6acac91a99856ac8c1819a20d271089086748fcc85c69a"} Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.779303 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebd9db4d3bfd046cfc6acac91a99856ac8c1819a20d271089086748fcc85c69a" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.779318 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-9nqzd" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.873562 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-r7zn5"] Nov 24 08:38:11 crc kubenswrapper[4799]: E1124 08:38:11.874404 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" containerName="extract-utilities" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.874532 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" containerName="extract-utilities" Nov 24 08:38:11 crc kubenswrapper[4799]: E1124 08:38:11.874617 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80d7525-1ef0-4824-857a-5264a31f9b73" containerName="install-os-openstack-openstack-cell1" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.874682 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80d7525-1ef0-4824-857a-5264a31f9b73" containerName="install-os-openstack-openstack-cell1" Nov 24 08:38:11 crc kubenswrapper[4799]: E1124 08:38:11.874748 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" containerName="extract-content" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.874811 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" containerName="extract-content" Nov 24 08:38:11 crc kubenswrapper[4799]: E1124 08:38:11.874917 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" containerName="registry-server" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.874996 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" containerName="registry-server" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.875352 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcb322a-ed70-4d00-b5b6-a3cfe1e93343" containerName="registry-server" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.875451 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80d7525-1ef0-4824-857a-5264a31f9b73" containerName="install-os-openstack-openstack-cell1" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.876466 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.879526 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.879651 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.880083 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.890335 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:38:11 crc kubenswrapper[4799]: I1124 08:38:11.906905 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-r7zn5"] Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.045271 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ceph\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.045382 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-inventory\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.046152 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ssh-key\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.046532 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvfsv\" (UniqueName: \"kubernetes.io/projected/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-kube-api-access-vvfsv\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.148594 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvfsv\" (UniqueName: \"kubernetes.io/projected/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-kube-api-access-vvfsv\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.148708 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ceph\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.148777 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-inventory\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.148819 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ssh-key\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.152544 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ssh-key\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.153242 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-inventory\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.157480 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ceph\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.169135 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvfsv\" (UniqueName: \"kubernetes.io/projected/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-kube-api-access-vvfsv\") pod \"configure-os-openstack-openstack-cell1-r7zn5\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.203954 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:12 crc kubenswrapper[4799]: I1124 08:38:12.917199 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-r7zn5"] Nov 24 08:38:13 crc kubenswrapper[4799]: I1124 08:38:13.800263 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" event={"ID":"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22","Type":"ContainerStarted","Data":"03259ccc2fd8e5991caf42276e03d4971d9c076c9edf7901de969d55e8bd2c7e"} Nov 24 08:38:13 crc kubenswrapper[4799]: I1124 08:38:13.800649 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" event={"ID":"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22","Type":"ContainerStarted","Data":"4149821d9f5173b69bc76d968e80a3e950e1f5229fb256ea9b56918ae1a4fa11"} Nov 24 08:38:13 crc kubenswrapper[4799]: I1124 08:38:13.828883 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" podStartSLOduration=2.433982147 podStartE2EDuration="2.828835735s" podCreationTimestamp="2025-11-24 08:38:11 +0000 UTC" firstStartedPulling="2025-11-24 08:38:12.912297407 +0000 UTC m=+6638.568279881" lastFinishedPulling="2025-11-24 08:38:13.307150995 +0000 UTC m=+6638.963133469" observedRunningTime="2025-11-24 08:38:13.81913836 +0000 UTC m=+6639.475120834" watchObservedRunningTime="2025-11-24 08:38:13.828835735 +0000 UTC m=+6639.484818219" Nov 24 08:38:18 crc kubenswrapper[4799]: I1124 08:38:18.628825 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:38:18 crc kubenswrapper[4799]: E1124 08:38:18.629566 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:38:32 crc kubenswrapper[4799]: I1124 08:38:32.629564 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:38:32 crc kubenswrapper[4799]: I1124 08:38:32.988407 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"d2fe628038dfea94f073cd1914b2280efdf1b72ad99513f5c1f470f890152657"} Nov 24 08:38:57 crc kubenswrapper[4799]: I1124 08:38:57.228134 4799 generic.go:334] "Generic (PLEG): container finished" podID="c29d6dd7-fd78-49e1-afcb-7b8349f6fe22" containerID="03259ccc2fd8e5991caf42276e03d4971d9c076c9edf7901de969d55e8bd2c7e" exitCode=0 Nov 24 08:38:57 crc kubenswrapper[4799]: I1124 08:38:57.228271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" event={"ID":"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22","Type":"ContainerDied","Data":"03259ccc2fd8e5991caf42276e03d4971d9c076c9edf7901de969d55e8bd2c7e"} Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.711065 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.827319 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ssh-key\") pod \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.827451 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ceph\") pod \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.827494 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvfsv\" (UniqueName: \"kubernetes.io/projected/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-kube-api-access-vvfsv\") pod \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.827573 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-inventory\") pod \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\" (UID: \"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22\") " Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.833232 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ceph" (OuterVolumeSpecName: "ceph") pod "c29d6dd7-fd78-49e1-afcb-7b8349f6fe22" (UID: "c29d6dd7-fd78-49e1-afcb-7b8349f6fe22"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.833630 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-kube-api-access-vvfsv" (OuterVolumeSpecName: "kube-api-access-vvfsv") pod "c29d6dd7-fd78-49e1-afcb-7b8349f6fe22" (UID: "c29d6dd7-fd78-49e1-afcb-7b8349f6fe22"). InnerVolumeSpecName "kube-api-access-vvfsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.857362 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c29d6dd7-fd78-49e1-afcb-7b8349f6fe22" (UID: "c29d6dd7-fd78-49e1-afcb-7b8349f6fe22"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.861148 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-inventory" (OuterVolumeSpecName: "inventory") pod "c29d6dd7-fd78-49e1-afcb-7b8349f6fe22" (UID: "c29d6dd7-fd78-49e1-afcb-7b8349f6fe22"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.933528 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.933558 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.933568 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvfsv\" (UniqueName: \"kubernetes.io/projected/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-kube-api-access-vvfsv\") on node \"crc\" DevicePath \"\"" Nov 24 08:38:58 crc kubenswrapper[4799]: I1124 08:38:58.933579 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c29d6dd7-fd78-49e1-afcb-7b8349f6fe22-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.256722 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" event={"ID":"c29d6dd7-fd78-49e1-afcb-7b8349f6fe22","Type":"ContainerDied","Data":"4149821d9f5173b69bc76d968e80a3e950e1f5229fb256ea9b56918ae1a4fa11"} Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.256760 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4149821d9f5173b69bc76d968e80a3e950e1f5229fb256ea9b56918ae1a4fa11" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.256841 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-r7zn5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.325997 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-pr2f5"] Nov 24 08:38:59 crc kubenswrapper[4799]: E1124 08:38:59.326421 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c29d6dd7-fd78-49e1-afcb-7b8349f6fe22" containerName="configure-os-openstack-openstack-cell1" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.326440 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c29d6dd7-fd78-49e1-afcb-7b8349f6fe22" containerName="configure-os-openstack-openstack-cell1" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.326622 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c29d6dd7-fd78-49e1-afcb-7b8349f6fe22" containerName="configure-os-openstack-openstack-cell1" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.327354 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.332646 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.332907 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.333046 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.333108 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.349779 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-pr2f5"] Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.444066 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.444164 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knxcg\" (UniqueName: \"kubernetes.io/projected/982af6fd-6fcf-42b7-ae60-1546609a2f8d-kube-api-access-knxcg\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.444916 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-inventory-0\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.444963 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ceph\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.546685 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.546800 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knxcg\" (UniqueName: \"kubernetes.io/projected/982af6fd-6fcf-42b7-ae60-1546609a2f8d-kube-api-access-knxcg\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.546896 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-inventory-0\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.546942 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ceph\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.552117 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ceph\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.552153 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-inventory-0\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.558268 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.565414 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knxcg\" (UniqueName: \"kubernetes.io/projected/982af6fd-6fcf-42b7-ae60-1546609a2f8d-kube-api-access-knxcg\") pod \"ssh-known-hosts-openstack-pr2f5\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:38:59 crc kubenswrapper[4799]: I1124 08:38:59.646295 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:39:00 crc kubenswrapper[4799]: I1124 08:39:00.201278 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-pr2f5"] Nov 24 08:39:00 crc kubenswrapper[4799]: I1124 08:39:00.283601 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-pr2f5" event={"ID":"982af6fd-6fcf-42b7-ae60-1546609a2f8d","Type":"ContainerStarted","Data":"49cacccffd5904179c42becd4fc3f0073a732230855e85daa5f7e0a7d62f12ab"} Nov 24 08:39:01 crc kubenswrapper[4799]: I1124 08:39:01.296378 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-pr2f5" event={"ID":"982af6fd-6fcf-42b7-ae60-1546609a2f8d","Type":"ContainerStarted","Data":"0bd5abb208a77e900443a949a92abfd44b71d3a1c469f7bcb843a044a48e7895"} Nov 24 08:39:01 crc kubenswrapper[4799]: I1124 08:39:01.322832 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-pr2f5" podStartSLOduration=1.9224142130000001 podStartE2EDuration="2.322747726s" podCreationTimestamp="2025-11-24 08:38:59 +0000 UTC" firstStartedPulling="2025-11-24 08:39:00.203416487 +0000 UTC m=+6685.859398961" lastFinishedPulling="2025-11-24 08:39:00.60375 +0000 UTC m=+6686.259732474" observedRunningTime="2025-11-24 08:39:01.310394855 +0000 UTC m=+6686.966377349" watchObservedRunningTime="2025-11-24 08:39:01.322747726 +0000 UTC m=+6686.978730210" Nov 24 08:39:10 crc kubenswrapper[4799]: I1124 08:39:10.395575 4799 generic.go:334] "Generic (PLEG): container finished" podID="982af6fd-6fcf-42b7-ae60-1546609a2f8d" containerID="0bd5abb208a77e900443a949a92abfd44b71d3a1c469f7bcb843a044a48e7895" exitCode=0 Nov 24 08:39:10 crc kubenswrapper[4799]: I1124 08:39:10.396247 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-pr2f5" event={"ID":"982af6fd-6fcf-42b7-ae60-1546609a2f8d","Type":"ContainerDied","Data":"0bd5abb208a77e900443a949a92abfd44b71d3a1c469f7bcb843a044a48e7895"} Nov 24 08:39:11 crc kubenswrapper[4799]: I1124 08:39:11.827338 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:39:11 crc kubenswrapper[4799]: I1124 08:39:11.928825 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knxcg\" (UniqueName: \"kubernetes.io/projected/982af6fd-6fcf-42b7-ae60-1546609a2f8d-kube-api-access-knxcg\") pod \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " Nov 24 08:39:11 crc kubenswrapper[4799]: I1124 08:39:11.928924 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ssh-key-openstack-cell1\") pod \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " Nov 24 08:39:11 crc kubenswrapper[4799]: I1124 08:39:11.929002 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-inventory-0\") pod \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " Nov 24 08:39:11 crc kubenswrapper[4799]: I1124 08:39:11.929029 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ceph\") pod \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\" (UID: \"982af6fd-6fcf-42b7-ae60-1546609a2f8d\") " Nov 24 08:39:11 crc kubenswrapper[4799]: I1124 08:39:11.935287 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/982af6fd-6fcf-42b7-ae60-1546609a2f8d-kube-api-access-knxcg" (OuterVolumeSpecName: "kube-api-access-knxcg") pod "982af6fd-6fcf-42b7-ae60-1546609a2f8d" (UID: "982af6fd-6fcf-42b7-ae60-1546609a2f8d"). InnerVolumeSpecName "kube-api-access-knxcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:39:11 crc kubenswrapper[4799]: I1124 08:39:11.936062 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ceph" (OuterVolumeSpecName: "ceph") pod "982af6fd-6fcf-42b7-ae60-1546609a2f8d" (UID: "982af6fd-6fcf-42b7-ae60-1546609a2f8d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:39:11 crc kubenswrapper[4799]: I1124 08:39:11.962460 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "982af6fd-6fcf-42b7-ae60-1546609a2f8d" (UID: "982af6fd-6fcf-42b7-ae60-1546609a2f8d"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:39:11 crc kubenswrapper[4799]: I1124 08:39:11.963019 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "982af6fd-6fcf-42b7-ae60-1546609a2f8d" (UID: "982af6fd-6fcf-42b7-ae60-1546609a2f8d"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.030270 4799 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.030310 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.030323 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knxcg\" (UniqueName: \"kubernetes.io/projected/982af6fd-6fcf-42b7-ae60-1546609a2f8d-kube-api-access-knxcg\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.030335 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/982af6fd-6fcf-42b7-ae60-1546609a2f8d-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.418595 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-pr2f5" event={"ID":"982af6fd-6fcf-42b7-ae60-1546609a2f8d","Type":"ContainerDied","Data":"49cacccffd5904179c42becd4fc3f0073a732230855e85daa5f7e0a7d62f12ab"} Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.418634 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49cacccffd5904179c42becd4fc3f0073a732230855e85daa5f7e0a7d62f12ab" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.418651 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-pr2f5" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.510574 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-l6s6d"] Nov 24 08:39:12 crc kubenswrapper[4799]: E1124 08:39:12.511016 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="982af6fd-6fcf-42b7-ae60-1546609a2f8d" containerName="ssh-known-hosts-openstack" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.511032 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="982af6fd-6fcf-42b7-ae60-1546609a2f8d" containerName="ssh-known-hosts-openstack" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.511247 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="982af6fd-6fcf-42b7-ae60-1546609a2f8d" containerName="ssh-known-hosts-openstack" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.511979 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.515931 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.516182 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.516211 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.520900 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.536363 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-l6s6d"] Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.542544 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-inventory\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.543066 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cvpt\" (UniqueName: \"kubernetes.io/projected/95317e48-8886-47e4-800e-a6b7e2663eb0-kube-api-access-4cvpt\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.543508 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ceph\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.543708 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ssh-key\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.645860 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-inventory\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.646026 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cvpt\" (UniqueName: \"kubernetes.io/projected/95317e48-8886-47e4-800e-a6b7e2663eb0-kube-api-access-4cvpt\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.646223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ceph\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.646315 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ssh-key\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.650136 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-inventory\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.650533 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ssh-key\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.652130 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ceph\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.670570 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cvpt\" (UniqueName: \"kubernetes.io/projected/95317e48-8886-47e4-800e-a6b7e2663eb0-kube-api-access-4cvpt\") pod \"run-os-openstack-openstack-cell1-l6s6d\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:12 crc kubenswrapper[4799]: I1124 08:39:12.830470 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:13 crc kubenswrapper[4799]: I1124 08:39:13.346340 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-l6s6d"] Nov 24 08:39:13 crc kubenswrapper[4799]: I1124 08:39:13.428682 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-l6s6d" event={"ID":"95317e48-8886-47e4-800e-a6b7e2663eb0","Type":"ContainerStarted","Data":"a532a4692f50554452e270cee53fa074f03850b33b347e76385bfe0ed1d7cf66"} Nov 24 08:39:14 crc kubenswrapper[4799]: I1124 08:39:14.438876 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-l6s6d" event={"ID":"95317e48-8886-47e4-800e-a6b7e2663eb0","Type":"ContainerStarted","Data":"6044e25eb41a1bdb7ca3c0926827d0045dc59c8e1886fa2c5a55ba49160ccdbf"} Nov 24 08:39:14 crc kubenswrapper[4799]: I1124 08:39:14.466977 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-l6s6d" podStartSLOduration=1.717345023 podStartE2EDuration="2.466956558s" podCreationTimestamp="2025-11-24 08:39:12 +0000 UTC" firstStartedPulling="2025-11-24 08:39:13.351440878 +0000 UTC m=+6699.007423352" lastFinishedPulling="2025-11-24 08:39:14.101052413 +0000 UTC m=+6699.757034887" observedRunningTime="2025-11-24 08:39:14.460208407 +0000 UTC m=+6700.116190881" watchObservedRunningTime="2025-11-24 08:39:14.466956558 +0000 UTC m=+6700.122939032" Nov 24 08:39:23 crc kubenswrapper[4799]: I1124 08:39:23.540357 4799 generic.go:334] "Generic (PLEG): container finished" podID="95317e48-8886-47e4-800e-a6b7e2663eb0" containerID="6044e25eb41a1bdb7ca3c0926827d0045dc59c8e1886fa2c5a55ba49160ccdbf" exitCode=0 Nov 24 08:39:23 crc kubenswrapper[4799]: I1124 08:39:23.540484 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-l6s6d" event={"ID":"95317e48-8886-47e4-800e-a6b7e2663eb0","Type":"ContainerDied","Data":"6044e25eb41a1bdb7ca3c0926827d0045dc59c8e1886fa2c5a55ba49160ccdbf"} Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.046352 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.226807 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ceph\") pod \"95317e48-8886-47e4-800e-a6b7e2663eb0\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.227213 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cvpt\" (UniqueName: \"kubernetes.io/projected/95317e48-8886-47e4-800e-a6b7e2663eb0-kube-api-access-4cvpt\") pod \"95317e48-8886-47e4-800e-a6b7e2663eb0\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.227255 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ssh-key\") pod \"95317e48-8886-47e4-800e-a6b7e2663eb0\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.227301 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-inventory\") pod \"95317e48-8886-47e4-800e-a6b7e2663eb0\" (UID: \"95317e48-8886-47e4-800e-a6b7e2663eb0\") " Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.233039 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ceph" (OuterVolumeSpecName: "ceph") pod "95317e48-8886-47e4-800e-a6b7e2663eb0" (UID: "95317e48-8886-47e4-800e-a6b7e2663eb0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.234128 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95317e48-8886-47e4-800e-a6b7e2663eb0-kube-api-access-4cvpt" (OuterVolumeSpecName: "kube-api-access-4cvpt") pod "95317e48-8886-47e4-800e-a6b7e2663eb0" (UID: "95317e48-8886-47e4-800e-a6b7e2663eb0"). InnerVolumeSpecName "kube-api-access-4cvpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.261536 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-inventory" (OuterVolumeSpecName: "inventory") pod "95317e48-8886-47e4-800e-a6b7e2663eb0" (UID: "95317e48-8886-47e4-800e-a6b7e2663eb0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.261998 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "95317e48-8886-47e4-800e-a6b7e2663eb0" (UID: "95317e48-8886-47e4-800e-a6b7e2663eb0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.332340 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.332391 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cvpt\" (UniqueName: \"kubernetes.io/projected/95317e48-8886-47e4-800e-a6b7e2663eb0-kube-api-access-4cvpt\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.332403 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.332412 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95317e48-8886-47e4-800e-a6b7e2663eb0-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.562638 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-l6s6d" event={"ID":"95317e48-8886-47e4-800e-a6b7e2663eb0","Type":"ContainerDied","Data":"a532a4692f50554452e270cee53fa074f03850b33b347e76385bfe0ed1d7cf66"} Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.562680 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a532a4692f50554452e270cee53fa074f03850b33b347e76385bfe0ed1d7cf66" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.562748 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-l6s6d" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.648897 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mpzh2"] Nov 24 08:39:25 crc kubenswrapper[4799]: E1124 08:39:25.649679 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95317e48-8886-47e4-800e-a6b7e2663eb0" containerName="run-os-openstack-openstack-cell1" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.649701 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="95317e48-8886-47e4-800e-a6b7e2663eb0" containerName="run-os-openstack-openstack-cell1" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.650083 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="95317e48-8886-47e4-800e-a6b7e2663eb0" containerName="run-os-openstack-openstack-cell1" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.651088 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.655402 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.655405 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.655557 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.657206 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.665690 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mpzh2"] Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.841979 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kkk2\" (UniqueName: \"kubernetes.io/projected/92a36210-a0e5-4548-99c1-0bf642f7aeb7-kube-api-access-6kkk2\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.842117 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.842149 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-inventory\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.843000 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ceph\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.945481 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kkk2\" (UniqueName: \"kubernetes.io/projected/92a36210-a0e5-4548-99c1-0bf642f7aeb7-kube-api-access-6kkk2\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.945631 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.945675 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-inventory\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.945710 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ceph\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.952038 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ceph\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.952469 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-inventory\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.965359 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kkk2\" (UniqueName: \"kubernetes.io/projected/92a36210-a0e5-4548-99c1-0bf642f7aeb7-kube-api-access-6kkk2\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.965387 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-mpzh2\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:25 crc kubenswrapper[4799]: I1124 08:39:25.973642 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:26 crc kubenswrapper[4799]: I1124 08:39:26.539841 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mpzh2"] Nov 24 08:39:26 crc kubenswrapper[4799]: I1124 08:39:26.581022 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" event={"ID":"92a36210-a0e5-4548-99c1-0bf642f7aeb7","Type":"ContainerStarted","Data":"fd73cf4607e105cef389658e7295ae8ed77b49dd8119205303e382a42ea0c0e7"} Nov 24 08:39:27 crc kubenswrapper[4799]: I1124 08:39:27.594755 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" event={"ID":"92a36210-a0e5-4548-99c1-0bf642f7aeb7","Type":"ContainerStarted","Data":"44a93eb9e551413ff1cf0cac782977dd9ab582f5c04b9b8ec601e93c1dcd0e13"} Nov 24 08:39:27 crc kubenswrapper[4799]: I1124 08:39:27.620463 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" podStartSLOduration=2.162412123 podStartE2EDuration="2.620423964s" podCreationTimestamp="2025-11-24 08:39:25 +0000 UTC" firstStartedPulling="2025-11-24 08:39:26.547931726 +0000 UTC m=+6712.203914210" lastFinishedPulling="2025-11-24 08:39:27.005943587 +0000 UTC m=+6712.661926051" observedRunningTime="2025-11-24 08:39:27.613501088 +0000 UTC m=+6713.269483562" watchObservedRunningTime="2025-11-24 08:39:27.620423964 +0000 UTC m=+6713.276406448" Nov 24 08:39:43 crc kubenswrapper[4799]: I1124 08:39:43.781801 4799 generic.go:334] "Generic (PLEG): container finished" podID="92a36210-a0e5-4548-99c1-0bf642f7aeb7" containerID="44a93eb9e551413ff1cf0cac782977dd9ab582f5c04b9b8ec601e93c1dcd0e13" exitCode=0 Nov 24 08:39:43 crc kubenswrapper[4799]: I1124 08:39:43.781897 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" event={"ID":"92a36210-a0e5-4548-99c1-0bf642f7aeb7","Type":"ContainerDied","Data":"44a93eb9e551413ff1cf0cac782977dd9ab582f5c04b9b8ec601e93c1dcd0e13"} Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.237895 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.366104 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kkk2\" (UniqueName: \"kubernetes.io/projected/92a36210-a0e5-4548-99c1-0bf642f7aeb7-kube-api-access-6kkk2\") pod \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.366173 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ceph\") pod \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.366242 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ssh-key\") pod \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.366391 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-inventory\") pod \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\" (UID: \"92a36210-a0e5-4548-99c1-0bf642f7aeb7\") " Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.375156 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ceph" (OuterVolumeSpecName: "ceph") pod "92a36210-a0e5-4548-99c1-0bf642f7aeb7" (UID: "92a36210-a0e5-4548-99c1-0bf642f7aeb7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.375187 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a36210-a0e5-4548-99c1-0bf642f7aeb7-kube-api-access-6kkk2" (OuterVolumeSpecName: "kube-api-access-6kkk2") pod "92a36210-a0e5-4548-99c1-0bf642f7aeb7" (UID: "92a36210-a0e5-4548-99c1-0bf642f7aeb7"). InnerVolumeSpecName "kube-api-access-6kkk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.397686 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-inventory" (OuterVolumeSpecName: "inventory") pod "92a36210-a0e5-4548-99c1-0bf642f7aeb7" (UID: "92a36210-a0e5-4548-99c1-0bf642f7aeb7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.397639 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "92a36210-a0e5-4548-99c1-0bf642f7aeb7" (UID: "92a36210-a0e5-4548-99c1-0bf642f7aeb7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.469098 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kkk2\" (UniqueName: \"kubernetes.io/projected/92a36210-a0e5-4548-99c1-0bf642f7aeb7-kube-api-access-6kkk2\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.469329 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.469449 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.469518 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92a36210-a0e5-4548-99c1-0bf642f7aeb7-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.803186 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" event={"ID":"92a36210-a0e5-4548-99c1-0bf642f7aeb7","Type":"ContainerDied","Data":"fd73cf4607e105cef389658e7295ae8ed77b49dd8119205303e382a42ea0c0e7"} Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.803592 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd73cf4607e105cef389658e7295ae8ed77b49dd8119205303e382a42ea0c0e7" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.803238 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mpzh2" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.898558 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-wzx69"] Nov 24 08:39:45 crc kubenswrapper[4799]: E1124 08:39:45.899333 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a36210-a0e5-4548-99c1-0bf642f7aeb7" containerName="reboot-os-openstack-openstack-cell1" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.899418 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a36210-a0e5-4548-99c1-0bf642f7aeb7" containerName="reboot-os-openstack-openstack-cell1" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.899724 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a36210-a0e5-4548-99c1-0bf642f7aeb7" containerName="reboot-os-openstack-openstack-cell1" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.900705 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.904535 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.905627 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.905805 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.906757 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-wzx69"] Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.912466 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.978870 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.978921 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.978991 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.979070 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-inventory\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.979103 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ssh-key\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.979148 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.979174 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.979193 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.979224 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ceph\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.979271 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5pfr\" (UniqueName: \"kubernetes.io/projected/b6c18c61-453f-4e00-ad20-6152e65992e7-kube-api-access-f5pfr\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.979352 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:45 crc kubenswrapper[4799]: I1124 08:39:45.979377 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081344 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081410 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081505 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081538 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081600 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081635 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-inventory\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081664 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ssh-key\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081706 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081735 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081753 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081783 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ceph\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.081831 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5pfr\" (UniqueName: \"kubernetes.io/projected/b6c18c61-453f-4e00-ad20-6152e65992e7-kube-api-access-f5pfr\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.089231 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.089290 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.089384 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ceph\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.089523 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.089789 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ssh-key\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.090861 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.091071 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.091650 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.092362 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.097958 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.101370 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-inventory\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.101962 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5pfr\" (UniqueName: \"kubernetes.io/projected/b6c18c61-453f-4e00-ad20-6152e65992e7-kube-api-access-f5pfr\") pod \"install-certs-openstack-openstack-cell1-wzx69\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.227965 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.778546 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-wzx69"] Nov 24 08:39:46 crc kubenswrapper[4799]: I1124 08:39:46.818200 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-wzx69" event={"ID":"b6c18c61-453f-4e00-ad20-6152e65992e7","Type":"ContainerStarted","Data":"71712f7e82f878159fa52c23ac44a6cc9b3f01fe264b0f13d26adbbe0fec2340"} Nov 24 08:39:47 crc kubenswrapper[4799]: I1124 08:39:47.830061 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-wzx69" event={"ID":"b6c18c61-453f-4e00-ad20-6152e65992e7","Type":"ContainerStarted","Data":"e476d80cd5ea051020dc1e11adf8fec1a6463ce14677fdd5272eab4f46522d7f"} Nov 24 08:39:47 crc kubenswrapper[4799]: I1124 08:39:47.862423 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-wzx69" podStartSLOduration=2.332642438 podStartE2EDuration="2.862404628s" podCreationTimestamp="2025-11-24 08:39:45 +0000 UTC" firstStartedPulling="2025-11-24 08:39:46.788667154 +0000 UTC m=+6732.444649628" lastFinishedPulling="2025-11-24 08:39:47.318429334 +0000 UTC m=+6732.974411818" observedRunningTime="2025-11-24 08:39:47.852454985 +0000 UTC m=+6733.508437479" watchObservedRunningTime="2025-11-24 08:39:47.862404628 +0000 UTC m=+6733.518387102" Nov 24 08:40:06 crc kubenswrapper[4799]: I1124 08:40:06.017161 4799 generic.go:334] "Generic (PLEG): container finished" podID="b6c18c61-453f-4e00-ad20-6152e65992e7" containerID="e476d80cd5ea051020dc1e11adf8fec1a6463ce14677fdd5272eab4f46522d7f" exitCode=0 Nov 24 08:40:06 crc kubenswrapper[4799]: I1124 08:40:06.017246 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-wzx69" event={"ID":"b6c18c61-453f-4e00-ad20-6152e65992e7","Type":"ContainerDied","Data":"e476d80cd5ea051020dc1e11adf8fec1a6463ce14677fdd5272eab4f46522d7f"} Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.548812 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.683548 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-bootstrap-combined-ca-bundle\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.683976 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-dhcp-combined-ca-bundle\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.684030 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-libvirt-combined-ca-bundle\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.684066 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ssh-key\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.684090 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ceph\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.684116 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-metadata-combined-ca-bundle\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.684143 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-telemetry-combined-ca-bundle\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.684219 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ovn-combined-ca-bundle\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.684240 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-sriov-combined-ca-bundle\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.684266 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-inventory\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.684316 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-nova-combined-ca-bundle\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.684426 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5pfr\" (UniqueName: \"kubernetes.io/projected/b6c18c61-453f-4e00-ad20-6152e65992e7-kube-api-access-f5pfr\") pod \"b6c18c61-453f-4e00-ad20-6152e65992e7\" (UID: \"b6c18c61-453f-4e00-ad20-6152e65992e7\") " Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.691233 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.691652 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.693575 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.694220 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.694340 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.694450 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6c18c61-453f-4e00-ad20-6152e65992e7-kube-api-access-f5pfr" (OuterVolumeSpecName: "kube-api-access-f5pfr") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "kube-api-access-f5pfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.696992 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ceph" (OuterVolumeSpecName: "ceph") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.700488 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.700628 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.703308 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.721226 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-inventory" (OuterVolumeSpecName: "inventory") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.734249 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b6c18c61-453f-4e00-ad20-6152e65992e7" (UID: "b6c18c61-453f-4e00-ad20-6152e65992e7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790008 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790067 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790084 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790095 4799 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790113 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5pfr\" (UniqueName: \"kubernetes.io/projected/b6c18c61-453f-4e00-ad20-6152e65992e7-kube-api-access-f5pfr\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790132 4799 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790148 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790158 4799 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790168 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790176 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790185 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:07 crc kubenswrapper[4799]: I1124 08:40:07.790194 4799 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c18c61-453f-4e00-ad20-6152e65992e7-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.046365 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-wzx69" event={"ID":"b6c18c61-453f-4e00-ad20-6152e65992e7","Type":"ContainerDied","Data":"71712f7e82f878159fa52c23ac44a6cc9b3f01fe264b0f13d26adbbe0fec2340"} Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.049756 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71712f7e82f878159fa52c23ac44a6cc9b3f01fe264b0f13d26adbbe0fec2340" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.050141 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-wzx69" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.223283 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-7ddgn"] Nov 24 08:40:08 crc kubenswrapper[4799]: E1124 08:40:08.223716 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c18c61-453f-4e00-ad20-6152e65992e7" containerName="install-certs-openstack-openstack-cell1" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.223732 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c18c61-453f-4e00-ad20-6152e65992e7" containerName="install-certs-openstack-openstack-cell1" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.223999 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6c18c61-453f-4e00-ad20-6152e65992e7" containerName="install-certs-openstack-openstack-cell1" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.224957 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.227519 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.228168 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.230206 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.231808 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.237365 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-7ddgn"] Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.301695 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ceph\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.301760 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s78kd\" (UniqueName: \"kubernetes.io/projected/2907da3b-efc0-488e-b8bd-6b0e4940a868-kube-api-access-s78kd\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.302137 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-inventory\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.302364 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.404723 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-inventory\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.404856 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.404968 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ceph\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.405019 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s78kd\" (UniqueName: \"kubernetes.io/projected/2907da3b-efc0-488e-b8bd-6b0e4940a868-kube-api-access-s78kd\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.408464 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-inventory\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.408812 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.408974 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ceph\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.423681 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s78kd\" (UniqueName: \"kubernetes.io/projected/2907da3b-efc0-488e-b8bd-6b0e4940a868-kube-api-access-s78kd\") pod \"ceph-client-openstack-openstack-cell1-7ddgn\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:08 crc kubenswrapper[4799]: I1124 08:40:08.545326 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:09 crc kubenswrapper[4799]: I1124 08:40:09.076618 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-7ddgn"] Nov 24 08:40:10 crc kubenswrapper[4799]: I1124 08:40:10.067474 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" event={"ID":"2907da3b-efc0-488e-b8bd-6b0e4940a868","Type":"ContainerStarted","Data":"3bdcbcafedabb5564184ccbefa7b149ac08cdcb89b1eee07e623271073815450"} Nov 24 08:40:10 crc kubenswrapper[4799]: I1124 08:40:10.067894 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" event={"ID":"2907da3b-efc0-488e-b8bd-6b0e4940a868","Type":"ContainerStarted","Data":"aa6c9f821f2e8f969983ae077fc2a13661b4bf8ef6de79e61b727c688f6fecb6"} Nov 24 08:40:10 crc kubenswrapper[4799]: I1124 08:40:10.095529 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" podStartSLOduration=1.635121119 podStartE2EDuration="2.095504568s" podCreationTimestamp="2025-11-24 08:40:08 +0000 UTC" firstStartedPulling="2025-11-24 08:40:09.080855462 +0000 UTC m=+6754.736837946" lastFinishedPulling="2025-11-24 08:40:09.541238921 +0000 UTC m=+6755.197221395" observedRunningTime="2025-11-24 08:40:10.083368433 +0000 UTC m=+6755.739350907" watchObservedRunningTime="2025-11-24 08:40:10.095504568 +0000 UTC m=+6755.751487042" Nov 24 08:40:15 crc kubenswrapper[4799]: I1124 08:40:15.129919 4799 generic.go:334] "Generic (PLEG): container finished" podID="2907da3b-efc0-488e-b8bd-6b0e4940a868" containerID="3bdcbcafedabb5564184ccbefa7b149ac08cdcb89b1eee07e623271073815450" exitCode=0 Nov 24 08:40:15 crc kubenswrapper[4799]: I1124 08:40:15.130068 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" event={"ID":"2907da3b-efc0-488e-b8bd-6b0e4940a868","Type":"ContainerDied","Data":"3bdcbcafedabb5564184ccbefa7b149ac08cdcb89b1eee07e623271073815450"} Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.618832 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.680242 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ceph\") pod \"2907da3b-efc0-488e-b8bd-6b0e4940a868\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.680372 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ssh-key\") pod \"2907da3b-efc0-488e-b8bd-6b0e4940a868\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.680505 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-inventory\") pod \"2907da3b-efc0-488e-b8bd-6b0e4940a868\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.680551 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s78kd\" (UniqueName: \"kubernetes.io/projected/2907da3b-efc0-488e-b8bd-6b0e4940a868-kube-api-access-s78kd\") pod \"2907da3b-efc0-488e-b8bd-6b0e4940a868\" (UID: \"2907da3b-efc0-488e-b8bd-6b0e4940a868\") " Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.688045 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ceph" (OuterVolumeSpecName: "ceph") pod "2907da3b-efc0-488e-b8bd-6b0e4940a868" (UID: "2907da3b-efc0-488e-b8bd-6b0e4940a868"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.688506 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2907da3b-efc0-488e-b8bd-6b0e4940a868-kube-api-access-s78kd" (OuterVolumeSpecName: "kube-api-access-s78kd") pod "2907da3b-efc0-488e-b8bd-6b0e4940a868" (UID: "2907da3b-efc0-488e-b8bd-6b0e4940a868"). InnerVolumeSpecName "kube-api-access-s78kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.712489 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-inventory" (OuterVolumeSpecName: "inventory") pod "2907da3b-efc0-488e-b8bd-6b0e4940a868" (UID: "2907da3b-efc0-488e-b8bd-6b0e4940a868"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.716064 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2907da3b-efc0-488e-b8bd-6b0e4940a868" (UID: "2907da3b-efc0-488e-b8bd-6b0e4940a868"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.784580 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.784632 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s78kd\" (UniqueName: \"kubernetes.io/projected/2907da3b-efc0-488e-b8bd-6b0e4940a868-kube-api-access-s78kd\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.784647 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:16 crc kubenswrapper[4799]: I1124 08:40:16.784661 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2907da3b-efc0-488e-b8bd-6b0e4940a868-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.153045 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" event={"ID":"2907da3b-efc0-488e-b8bd-6b0e4940a868","Type":"ContainerDied","Data":"aa6c9f821f2e8f969983ae077fc2a13661b4bf8ef6de79e61b727c688f6fecb6"} Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.153107 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa6c9f821f2e8f969983ae077fc2a13661b4bf8ef6de79e61b727c688f6fecb6" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.153182 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-7ddgn" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.329330 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-7h9f4"] Nov 24 08:40:17 crc kubenswrapper[4799]: E1124 08:40:17.330439 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2907da3b-efc0-488e-b8bd-6b0e4940a868" containerName="ceph-client-openstack-openstack-cell1" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.330547 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2907da3b-efc0-488e-b8bd-6b0e4940a868" containerName="ceph-client-openstack-openstack-cell1" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.330923 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2907da3b-efc0-488e-b8bd-6b0e4940a868" containerName="ceph-client-openstack-openstack-cell1" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.332176 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.334525 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.335176 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.335426 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.335655 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.336236 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.341436 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-7h9f4"] Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.402775 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcsgv\" (UniqueName: \"kubernetes.io/projected/93716fdc-7850-4f86-af66-6511c476f8b3-kube-api-access-gcsgv\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.402878 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ceph\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.403051 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/93716fdc-7850-4f86-af66-6511c476f8b3-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.403141 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-inventory\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.403176 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.403237 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ssh-key\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.505114 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ssh-key\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.505507 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcsgv\" (UniqueName: \"kubernetes.io/projected/93716fdc-7850-4f86-af66-6511c476f8b3-kube-api-access-gcsgv\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.505632 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ceph\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.505842 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/93716fdc-7850-4f86-af66-6511c476f8b3-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.506004 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-inventory\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.506091 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.507408 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/93716fdc-7850-4f86-af66-6511c476f8b3-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.511055 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ceph\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.511435 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-inventory\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.512726 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ssh-key\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.513268 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.529576 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcsgv\" (UniqueName: \"kubernetes.io/projected/93716fdc-7850-4f86-af66-6511c476f8b3-kube-api-access-gcsgv\") pod \"ovn-openstack-openstack-cell1-7h9f4\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:17 crc kubenswrapper[4799]: I1124 08:40:17.664357 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:40:18 crc kubenswrapper[4799]: I1124 08:40:18.062994 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-7h9f4"] Nov 24 08:40:18 crc kubenswrapper[4799]: W1124 08:40:18.088149 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93716fdc_7850_4f86_af66_6511c476f8b3.slice/crio-a6ecb69c280b99070553aeffe3755d5cec246233e0659aa5027cddf24d26fa6e WatchSource:0}: Error finding container a6ecb69c280b99070553aeffe3755d5cec246233e0659aa5027cddf24d26fa6e: Status 404 returned error can't find the container with id a6ecb69c280b99070553aeffe3755d5cec246233e0659aa5027cddf24d26fa6e Nov 24 08:40:18 crc kubenswrapper[4799]: I1124 08:40:18.165074 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-7h9f4" event={"ID":"93716fdc-7850-4f86-af66-6511c476f8b3","Type":"ContainerStarted","Data":"a6ecb69c280b99070553aeffe3755d5cec246233e0659aa5027cddf24d26fa6e"} Nov 24 08:40:22 crc kubenswrapper[4799]: I1124 08:40:22.217529 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-7h9f4" event={"ID":"93716fdc-7850-4f86-af66-6511c476f8b3","Type":"ContainerStarted","Data":"5f1e2334d826b020a42f3462610e33f97921491f4ac56c0a2686800d1759a83c"} Nov 24 08:40:22 crc kubenswrapper[4799]: I1124 08:40:22.243897 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-7h9f4" podStartSLOduration=1.745054381 podStartE2EDuration="5.243870289s" podCreationTimestamp="2025-11-24 08:40:17 +0000 UTC" firstStartedPulling="2025-11-24 08:40:18.094209041 +0000 UTC m=+6763.750191515" lastFinishedPulling="2025-11-24 08:40:21.593024949 +0000 UTC m=+6767.249007423" observedRunningTime="2025-11-24 08:40:22.233286229 +0000 UTC m=+6767.889268713" watchObservedRunningTime="2025-11-24 08:40:22.243870289 +0000 UTC m=+6767.899852783" Nov 24 08:40:50 crc kubenswrapper[4799]: I1124 08:40:50.400660 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:40:50 crc kubenswrapper[4799]: I1124 08:40:50.401260 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:41:20 crc kubenswrapper[4799]: I1124 08:41:20.400929 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:41:20 crc kubenswrapper[4799]: I1124 08:41:20.401567 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:41:25 crc kubenswrapper[4799]: I1124 08:41:25.914169 4799 generic.go:334] "Generic (PLEG): container finished" podID="93716fdc-7850-4f86-af66-6511c476f8b3" containerID="5f1e2334d826b020a42f3462610e33f97921491f4ac56c0a2686800d1759a83c" exitCode=0 Nov 24 08:41:25 crc kubenswrapper[4799]: I1124 08:41:25.914265 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-7h9f4" event={"ID":"93716fdc-7850-4f86-af66-6511c476f8b3","Type":"ContainerDied","Data":"5f1e2334d826b020a42f3462610e33f97921491f4ac56c0a2686800d1759a83c"} Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.352985 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.424605 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ovn-combined-ca-bundle\") pod \"93716fdc-7850-4f86-af66-6511c476f8b3\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.424646 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ceph\") pod \"93716fdc-7850-4f86-af66-6511c476f8b3\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.424716 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ssh-key\") pod \"93716fdc-7850-4f86-af66-6511c476f8b3\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.424787 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcsgv\" (UniqueName: \"kubernetes.io/projected/93716fdc-7850-4f86-af66-6511c476f8b3-kube-api-access-gcsgv\") pod \"93716fdc-7850-4f86-af66-6511c476f8b3\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.425023 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/93716fdc-7850-4f86-af66-6511c476f8b3-ovncontroller-config-0\") pod \"93716fdc-7850-4f86-af66-6511c476f8b3\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.425076 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-inventory\") pod \"93716fdc-7850-4f86-af66-6511c476f8b3\" (UID: \"93716fdc-7850-4f86-af66-6511c476f8b3\") " Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.432083 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93716fdc-7850-4f86-af66-6511c476f8b3-kube-api-access-gcsgv" (OuterVolumeSpecName: "kube-api-access-gcsgv") pod "93716fdc-7850-4f86-af66-6511c476f8b3" (UID: "93716fdc-7850-4f86-af66-6511c476f8b3"). InnerVolumeSpecName "kube-api-access-gcsgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.432940 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ceph" (OuterVolumeSpecName: "ceph") pod "93716fdc-7850-4f86-af66-6511c476f8b3" (UID: "93716fdc-7850-4f86-af66-6511c476f8b3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.445992 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "93716fdc-7850-4f86-af66-6511c476f8b3" (UID: "93716fdc-7850-4f86-af66-6511c476f8b3"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.470205 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "93716fdc-7850-4f86-af66-6511c476f8b3" (UID: "93716fdc-7850-4f86-af66-6511c476f8b3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.472210 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-inventory" (OuterVolumeSpecName: "inventory") pod "93716fdc-7850-4f86-af66-6511c476f8b3" (UID: "93716fdc-7850-4f86-af66-6511c476f8b3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.480446 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93716fdc-7850-4f86-af66-6511c476f8b3-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "93716fdc-7850-4f86-af66-6511c476f8b3" (UID: "93716fdc-7850-4f86-af66-6511c476f8b3"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.528017 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.528054 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.528064 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.528072 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93716fdc-7850-4f86-af66-6511c476f8b3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.528082 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcsgv\" (UniqueName: \"kubernetes.io/projected/93716fdc-7850-4f86-af66-6511c476f8b3-kube-api-access-gcsgv\") on node \"crc\" DevicePath \"\"" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.528090 4799 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/93716fdc-7850-4f86-af66-6511c476f8b3-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.936975 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-7h9f4" event={"ID":"93716fdc-7850-4f86-af66-6511c476f8b3","Type":"ContainerDied","Data":"a6ecb69c280b99070553aeffe3755d5cec246233e0659aa5027cddf24d26fa6e"} Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.937204 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6ecb69c280b99070553aeffe3755d5cec246233e0659aa5027cddf24d26fa6e" Nov 24 08:41:27 crc kubenswrapper[4799]: I1124 08:41:27.937027 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-7h9f4" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.027588 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-ts42m"] Nov 24 08:41:28 crc kubenswrapper[4799]: E1124 08:41:28.028104 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93716fdc-7850-4f86-af66-6511c476f8b3" containerName="ovn-openstack-openstack-cell1" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.028125 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="93716fdc-7850-4f86-af66-6511c476f8b3" containerName="ovn-openstack-openstack-cell1" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.028365 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="93716fdc-7850-4f86-af66-6511c476f8b3" containerName="ovn-openstack-openstack-cell1" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.029178 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.033396 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.033556 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.033776 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.033925 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.033954 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.040561 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.047279 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-ts42m"] Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.141471 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.141532 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.141567 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.141723 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpwhb\" (UniqueName: \"kubernetes.io/projected/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-kube-api-access-kpwhb\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.141784 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.141824 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.141941 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.243331 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.243431 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.243495 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.243542 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.243585 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.243651 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpwhb\" (UniqueName: \"kubernetes.io/projected/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-kube-api-access-kpwhb\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.243704 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.248212 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.248356 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.248456 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.250571 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.258241 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.258669 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.266033 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpwhb\" (UniqueName: \"kubernetes.io/projected/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-kube-api-access-kpwhb\") pod \"neutron-metadata-openstack-openstack-cell1-ts42m\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.351673 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.928176 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-ts42m"] Nov 24 08:41:28 crc kubenswrapper[4799]: I1124 08:41:28.948161 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" event={"ID":"4b9803e3-b71b-4cb3-a8ec-098180db8e5c","Type":"ContainerStarted","Data":"86c9a47908dc8e8d1f23d80af881cb734557ffe3970fc1fa5a8252ed1b0fcae5"} Nov 24 08:41:29 crc kubenswrapper[4799]: I1124 08:41:29.965283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" event={"ID":"4b9803e3-b71b-4cb3-a8ec-098180db8e5c","Type":"ContainerStarted","Data":"242d145b5da1e02e621047d85ed94c3acca32b9368338cc7c6f17e7ffeac5abd"} Nov 24 08:41:29 crc kubenswrapper[4799]: I1124 08:41:29.986508 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" podStartSLOduration=1.435736256 podStartE2EDuration="1.986487008s" podCreationTimestamp="2025-11-24 08:41:28 +0000 UTC" firstStartedPulling="2025-11-24 08:41:28.929787534 +0000 UTC m=+6834.585770008" lastFinishedPulling="2025-11-24 08:41:29.480538296 +0000 UTC m=+6835.136520760" observedRunningTime="2025-11-24 08:41:29.981476386 +0000 UTC m=+6835.637458860" watchObservedRunningTime="2025-11-24 08:41:29.986487008 +0000 UTC m=+6835.642469482" Nov 24 08:41:50 crc kubenswrapper[4799]: I1124 08:41:50.400932 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:41:50 crc kubenswrapper[4799]: I1124 08:41:50.401570 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:41:50 crc kubenswrapper[4799]: I1124 08:41:50.401626 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 08:41:50 crc kubenswrapper[4799]: I1124 08:41:50.402825 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d2fe628038dfea94f073cd1914b2280efdf1b72ad99513f5c1f470f890152657"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:41:50 crc kubenswrapper[4799]: I1124 08:41:50.402924 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://d2fe628038dfea94f073cd1914b2280efdf1b72ad99513f5c1f470f890152657" gracePeriod=600 Nov 24 08:41:51 crc kubenswrapper[4799]: I1124 08:41:51.184299 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="d2fe628038dfea94f073cd1914b2280efdf1b72ad99513f5c1f470f890152657" exitCode=0 Nov 24 08:41:51 crc kubenswrapper[4799]: I1124 08:41:51.184381 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"d2fe628038dfea94f073cd1914b2280efdf1b72ad99513f5c1f470f890152657"} Nov 24 08:41:51 crc kubenswrapper[4799]: I1124 08:41:51.184973 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953"} Nov 24 08:41:51 crc kubenswrapper[4799]: I1124 08:41:51.185010 4799 scope.go:117] "RemoveContainer" containerID="5301834e706d99d8d6d80ecc90941d002bfe4288348c3458b3a434cc8d1c0e5b" Nov 24 08:42:20 crc kubenswrapper[4799]: I1124 08:42:20.486752 4799 generic.go:334] "Generic (PLEG): container finished" podID="4b9803e3-b71b-4cb3-a8ec-098180db8e5c" containerID="242d145b5da1e02e621047d85ed94c3acca32b9368338cc7c6f17e7ffeac5abd" exitCode=0 Nov 24 08:42:20 crc kubenswrapper[4799]: I1124 08:42:20.486904 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" event={"ID":"4b9803e3-b71b-4cb3-a8ec-098180db8e5c","Type":"ContainerDied","Data":"242d145b5da1e02e621047d85ed94c3acca32b9368338cc7c6f17e7ffeac5abd"} Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.047363 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.199500 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpwhb\" (UniqueName: \"kubernetes.io/projected/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-kube-api-access-kpwhb\") pod \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.199593 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-metadata-combined-ca-bundle\") pod \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.199641 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ceph\") pod \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.199691 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ssh-key\") pod \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.199724 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-inventory\") pod \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.199783 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.200031 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-nova-metadata-neutron-config-0\") pod \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\" (UID: \"4b9803e3-b71b-4cb3-a8ec-098180db8e5c\") " Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.206181 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ceph" (OuterVolumeSpecName: "ceph") pod "4b9803e3-b71b-4cb3-a8ec-098180db8e5c" (UID: "4b9803e3-b71b-4cb3-a8ec-098180db8e5c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.206316 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-kube-api-access-kpwhb" (OuterVolumeSpecName: "kube-api-access-kpwhb") pod "4b9803e3-b71b-4cb3-a8ec-098180db8e5c" (UID: "4b9803e3-b71b-4cb3-a8ec-098180db8e5c"). InnerVolumeSpecName "kube-api-access-kpwhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.217057 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4b9803e3-b71b-4cb3-a8ec-098180db8e5c" (UID: "4b9803e3-b71b-4cb3-a8ec-098180db8e5c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.231708 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "4b9803e3-b71b-4cb3-a8ec-098180db8e5c" (UID: "4b9803e3-b71b-4cb3-a8ec-098180db8e5c"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.232074 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b9803e3-b71b-4cb3-a8ec-098180db8e5c" (UID: "4b9803e3-b71b-4cb3-a8ec-098180db8e5c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.241075 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-inventory" (OuterVolumeSpecName: "inventory") pod "4b9803e3-b71b-4cb3-a8ec-098180db8e5c" (UID: "4b9803e3-b71b-4cb3-a8ec-098180db8e5c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.243072 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "4b9803e3-b71b-4cb3-a8ec-098180db8e5c" (UID: "4b9803e3-b71b-4cb3-a8ec-098180db8e5c"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.301909 4799 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.302572 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpwhb\" (UniqueName: \"kubernetes.io/projected/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-kube-api-access-kpwhb\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.302663 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.302732 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.302793 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.302859 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.302943 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b9803e3-b71b-4cb3-a8ec-098180db8e5c-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.509360 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" event={"ID":"4b9803e3-b71b-4cb3-a8ec-098180db8e5c","Type":"ContainerDied","Data":"86c9a47908dc8e8d1f23d80af881cb734557ffe3970fc1fa5a8252ed1b0fcae5"} Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.509592 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86c9a47908dc8e8d1f23d80af881cb734557ffe3970fc1fa5a8252ed1b0fcae5" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.509751 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-ts42m" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.617718 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-nnrmm"] Nov 24 08:42:22 crc kubenswrapper[4799]: E1124 08:42:22.618225 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b9803e3-b71b-4cb3-a8ec-098180db8e5c" containerName="neutron-metadata-openstack-openstack-cell1" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.618247 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b9803e3-b71b-4cb3-a8ec-098180db8e5c" containerName="neutron-metadata-openstack-openstack-cell1" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.618536 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b9803e3-b71b-4cb3-a8ec-098180db8e5c" containerName="neutron-metadata-openstack-openstack-cell1" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.619374 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.623789 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.624757 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.625080 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.627650 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.631396 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-nnrmm"] Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.633471 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.710942 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-inventory\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.710986 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.711151 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnm6j\" (UniqueName: \"kubernetes.io/projected/2ebcd170-a63a-4355-9046-17e8f42f4dc8-kube-api-access-vnm6j\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.711184 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ceph\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.711204 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.711223 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ssh-key\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.814360 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ceph\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.814829 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.815049 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ssh-key\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.815293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-inventory\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.815439 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.815758 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnm6j\" (UniqueName: \"kubernetes.io/projected/2ebcd170-a63a-4355-9046-17e8f42f4dc8-kube-api-access-vnm6j\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.819527 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ceph\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.820133 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.820583 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ssh-key\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.821677 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-inventory\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.832926 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.837940 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnm6j\" (UniqueName: \"kubernetes.io/projected/2ebcd170-a63a-4355-9046-17e8f42f4dc8-kube-api-access-vnm6j\") pod \"libvirt-openstack-openstack-cell1-nnrmm\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:22 crc kubenswrapper[4799]: I1124 08:42:22.937849 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:42:23 crc kubenswrapper[4799]: I1124 08:42:23.492298 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-nnrmm"] Nov 24 08:42:23 crc kubenswrapper[4799]: W1124 08:42:23.522543 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ebcd170_a63a_4355_9046_17e8f42f4dc8.slice/crio-230a14ea43994626f614382d595480944f1b008138b9a08ec11dded4ac3728b5 WatchSource:0}: Error finding container 230a14ea43994626f614382d595480944f1b008138b9a08ec11dded4ac3728b5: Status 404 returned error can't find the container with id 230a14ea43994626f614382d595480944f1b008138b9a08ec11dded4ac3728b5 Nov 24 08:42:23 crc kubenswrapper[4799]: I1124 08:42:23.525370 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:42:24 crc kubenswrapper[4799]: I1124 08:42:24.539942 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" event={"ID":"2ebcd170-a63a-4355-9046-17e8f42f4dc8","Type":"ContainerStarted","Data":"8ce01d3eb969032db337bb3c5e2bed09004ec025e4cf57e88b8fde35788fe264"} Nov 24 08:42:24 crc kubenswrapper[4799]: I1124 08:42:24.540492 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" event={"ID":"2ebcd170-a63a-4355-9046-17e8f42f4dc8","Type":"ContainerStarted","Data":"230a14ea43994626f614382d595480944f1b008138b9a08ec11dded4ac3728b5"} Nov 24 08:42:24 crc kubenswrapper[4799]: I1124 08:42:24.562975 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" podStartSLOduration=1.988022527 podStartE2EDuration="2.562955317s" podCreationTimestamp="2025-11-24 08:42:22 +0000 UTC" firstStartedPulling="2025-11-24 08:42:23.524972033 +0000 UTC m=+6889.180954507" lastFinishedPulling="2025-11-24 08:42:24.099904823 +0000 UTC m=+6889.755887297" observedRunningTime="2025-11-24 08:42:24.554118286 +0000 UTC m=+6890.210100760" watchObservedRunningTime="2025-11-24 08:42:24.562955317 +0000 UTC m=+6890.218937791" Nov 24 08:43:50 crc kubenswrapper[4799]: I1124 08:43:50.400451 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:43:50 crc kubenswrapper[4799]: I1124 08:43:50.400958 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:44:20 crc kubenswrapper[4799]: I1124 08:44:20.401185 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:44:20 crc kubenswrapper[4799]: I1124 08:44:20.401792 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:44:38 crc kubenswrapper[4799]: I1124 08:44:38.960920 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zzksx"] Nov 24 08:44:38 crc kubenswrapper[4799]: I1124 08:44:38.965760 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:38 crc kubenswrapper[4799]: I1124 08:44:38.984320 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zzksx"] Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.124608 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkmwr\" (UniqueName: \"kubernetes.io/projected/331f65db-6ad5-4d2e-8af9-b0865ba609a7-kube-api-access-pkmwr\") pod \"certified-operators-zzksx\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.124723 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-catalog-content\") pod \"certified-operators-zzksx\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.125571 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-utilities\") pod \"certified-operators-zzksx\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.228215 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkmwr\" (UniqueName: \"kubernetes.io/projected/331f65db-6ad5-4d2e-8af9-b0865ba609a7-kube-api-access-pkmwr\") pod \"certified-operators-zzksx\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.228264 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-catalog-content\") pod \"certified-operators-zzksx\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.228445 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-utilities\") pod \"certified-operators-zzksx\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.229009 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-utilities\") pod \"certified-operators-zzksx\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.229277 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-catalog-content\") pod \"certified-operators-zzksx\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.254212 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkmwr\" (UniqueName: \"kubernetes.io/projected/331f65db-6ad5-4d2e-8af9-b0865ba609a7-kube-api-access-pkmwr\") pod \"certified-operators-zzksx\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.322231 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.890554 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zzksx"] Nov 24 08:44:39 crc kubenswrapper[4799]: I1124 08:44:39.948355 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzksx" event={"ID":"331f65db-6ad5-4d2e-8af9-b0865ba609a7","Type":"ContainerStarted","Data":"39b719bc64e6653cdb312c028c2f00a781ebed0906872c90f02e1fc0ee3aa5e3"} Nov 24 08:44:40 crc kubenswrapper[4799]: I1124 08:44:40.961105 4799 generic.go:334] "Generic (PLEG): container finished" podID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" containerID="282d310fad3b4bc33f10a97f611a9d6f85615b1acecc103a202391b0a5a683f6" exitCode=0 Nov 24 08:44:40 crc kubenswrapper[4799]: I1124 08:44:40.961240 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzksx" event={"ID":"331f65db-6ad5-4d2e-8af9-b0865ba609a7","Type":"ContainerDied","Data":"282d310fad3b4bc33f10a97f611a9d6f85615b1acecc103a202391b0a5a683f6"} Nov 24 08:44:41 crc kubenswrapper[4799]: I1124 08:44:41.971166 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzksx" event={"ID":"331f65db-6ad5-4d2e-8af9-b0865ba609a7","Type":"ContainerStarted","Data":"aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d"} Nov 24 08:44:43 crc kubenswrapper[4799]: I1124 08:44:43.146024 4799 generic.go:334] "Generic (PLEG): container finished" podID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" containerID="aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d" exitCode=0 Nov 24 08:44:43 crc kubenswrapper[4799]: I1124 08:44:43.146404 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzksx" event={"ID":"331f65db-6ad5-4d2e-8af9-b0865ba609a7","Type":"ContainerDied","Data":"aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d"} Nov 24 08:44:44 crc kubenswrapper[4799]: I1124 08:44:44.158392 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzksx" event={"ID":"331f65db-6ad5-4d2e-8af9-b0865ba609a7","Type":"ContainerStarted","Data":"1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c"} Nov 24 08:44:44 crc kubenswrapper[4799]: I1124 08:44:44.182383 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zzksx" podStartSLOduration=3.440874205 podStartE2EDuration="6.182362625s" podCreationTimestamp="2025-11-24 08:44:38 +0000 UTC" firstStartedPulling="2025-11-24 08:44:40.964561895 +0000 UTC m=+7026.620544369" lastFinishedPulling="2025-11-24 08:44:43.706050305 +0000 UTC m=+7029.362032789" observedRunningTime="2025-11-24 08:44:44.174689068 +0000 UTC m=+7029.830671542" watchObservedRunningTime="2025-11-24 08:44:44.182362625 +0000 UTC m=+7029.838345099" Nov 24 08:44:49 crc kubenswrapper[4799]: I1124 08:44:49.322667 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:49 crc kubenswrapper[4799]: I1124 08:44:49.323751 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:49 crc kubenswrapper[4799]: I1124 08:44:49.381929 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:50 crc kubenswrapper[4799]: I1124 08:44:50.273219 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:50 crc kubenswrapper[4799]: I1124 08:44:50.338632 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zzksx"] Nov 24 08:44:50 crc kubenswrapper[4799]: I1124 08:44:50.400685 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:44:50 crc kubenswrapper[4799]: I1124 08:44:50.400748 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:44:50 crc kubenswrapper[4799]: I1124 08:44:50.400832 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 08:44:50 crc kubenswrapper[4799]: I1124 08:44:50.401687 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:44:50 crc kubenswrapper[4799]: I1124 08:44:50.401759 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" gracePeriod=600 Nov 24 08:44:50 crc kubenswrapper[4799]: E1124 08:44:50.539146 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:44:51 crc kubenswrapper[4799]: I1124 08:44:51.238800 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" exitCode=0 Nov 24 08:44:51 crc kubenswrapper[4799]: I1124 08:44:51.238898 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953"} Nov 24 08:44:51 crc kubenswrapper[4799]: I1124 08:44:51.238973 4799 scope.go:117] "RemoveContainer" containerID="d2fe628038dfea94f073cd1914b2280efdf1b72ad99513f5c1f470f890152657" Nov 24 08:44:51 crc kubenswrapper[4799]: I1124 08:44:51.240228 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:44:51 crc kubenswrapper[4799]: E1124 08:44:51.240556 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:44:52 crc kubenswrapper[4799]: I1124 08:44:52.253468 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zzksx" podUID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" containerName="registry-server" containerID="cri-o://1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c" gracePeriod=2 Nov 24 08:44:52 crc kubenswrapper[4799]: I1124 08:44:52.754382 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:52 crc kubenswrapper[4799]: I1124 08:44:52.852583 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-catalog-content\") pod \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " Nov 24 08:44:52 crc kubenswrapper[4799]: I1124 08:44:52.853051 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkmwr\" (UniqueName: \"kubernetes.io/projected/331f65db-6ad5-4d2e-8af9-b0865ba609a7-kube-api-access-pkmwr\") pod \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " Nov 24 08:44:52 crc kubenswrapper[4799]: I1124 08:44:52.853223 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-utilities\") pod \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\" (UID: \"331f65db-6ad5-4d2e-8af9-b0865ba609a7\") " Nov 24 08:44:52 crc kubenswrapper[4799]: I1124 08:44:52.854515 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-utilities" (OuterVolumeSpecName: "utilities") pod "331f65db-6ad5-4d2e-8af9-b0865ba609a7" (UID: "331f65db-6ad5-4d2e-8af9-b0865ba609a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:44:52 crc kubenswrapper[4799]: I1124 08:44:52.863601 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/331f65db-6ad5-4d2e-8af9-b0865ba609a7-kube-api-access-pkmwr" (OuterVolumeSpecName: "kube-api-access-pkmwr") pod "331f65db-6ad5-4d2e-8af9-b0865ba609a7" (UID: "331f65db-6ad5-4d2e-8af9-b0865ba609a7"). InnerVolumeSpecName "kube-api-access-pkmwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:44:52 crc kubenswrapper[4799]: I1124 08:44:52.899070 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "331f65db-6ad5-4d2e-8af9-b0865ba609a7" (UID: "331f65db-6ad5-4d2e-8af9-b0865ba609a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:44:52 crc kubenswrapper[4799]: I1124 08:44:52.956995 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:52 crc kubenswrapper[4799]: I1124 08:44:52.957056 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/331f65db-6ad5-4d2e-8af9-b0865ba609a7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:52 crc kubenswrapper[4799]: I1124 08:44:52.957071 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkmwr\" (UniqueName: \"kubernetes.io/projected/331f65db-6ad5-4d2e-8af9-b0865ba609a7-kube-api-access-pkmwr\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.265272 4799 generic.go:334] "Generic (PLEG): container finished" podID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" containerID="1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c" exitCode=0 Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.265319 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzksx" Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.265329 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzksx" event={"ID":"331f65db-6ad5-4d2e-8af9-b0865ba609a7","Type":"ContainerDied","Data":"1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c"} Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.265364 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzksx" event={"ID":"331f65db-6ad5-4d2e-8af9-b0865ba609a7","Type":"ContainerDied","Data":"39b719bc64e6653cdb312c028c2f00a781ebed0906872c90f02e1fc0ee3aa5e3"} Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.265388 4799 scope.go:117] "RemoveContainer" containerID="1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c" Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.305186 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zzksx"] Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.306983 4799 scope.go:117] "RemoveContainer" containerID="aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d" Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.319084 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zzksx"] Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.337126 4799 scope.go:117] "RemoveContainer" containerID="282d310fad3b4bc33f10a97f611a9d6f85615b1acecc103a202391b0a5a683f6" Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.373392 4799 scope.go:117] "RemoveContainer" containerID="1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c" Nov 24 08:44:53 crc kubenswrapper[4799]: E1124 08:44:53.373821 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c\": container with ID starting with 1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c not found: ID does not exist" containerID="1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c" Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.373913 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c"} err="failed to get container status \"1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c\": rpc error: code = NotFound desc = could not find container \"1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c\": container with ID starting with 1e0d6e849343088622bf86ea0f021074d0de6063ceda1202763176d302d9a59c not found: ID does not exist" Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.373940 4799 scope.go:117] "RemoveContainer" containerID="aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d" Nov 24 08:44:53 crc kubenswrapper[4799]: E1124 08:44:53.374307 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d\": container with ID starting with aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d not found: ID does not exist" containerID="aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d" Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.374336 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d"} err="failed to get container status \"aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d\": rpc error: code = NotFound desc = could not find container \"aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d\": container with ID starting with aebfdd887a794ecfc6e5e99bd05c6ff57fbf39250f6bc929688073413604205d not found: ID does not exist" Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.374354 4799 scope.go:117] "RemoveContainer" containerID="282d310fad3b4bc33f10a97f611a9d6f85615b1acecc103a202391b0a5a683f6" Nov 24 08:44:53 crc kubenswrapper[4799]: E1124 08:44:53.374782 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"282d310fad3b4bc33f10a97f611a9d6f85615b1acecc103a202391b0a5a683f6\": container with ID starting with 282d310fad3b4bc33f10a97f611a9d6f85615b1acecc103a202391b0a5a683f6 not found: ID does not exist" containerID="282d310fad3b4bc33f10a97f611a9d6f85615b1acecc103a202391b0a5a683f6" Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.374808 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"282d310fad3b4bc33f10a97f611a9d6f85615b1acecc103a202391b0a5a683f6"} err="failed to get container status \"282d310fad3b4bc33f10a97f611a9d6f85615b1acecc103a202391b0a5a683f6\": rpc error: code = NotFound desc = could not find container \"282d310fad3b4bc33f10a97f611a9d6f85615b1acecc103a202391b0a5a683f6\": container with ID starting with 282d310fad3b4bc33f10a97f611a9d6f85615b1acecc103a202391b0a5a683f6 not found: ID does not exist" Nov 24 08:44:53 crc kubenswrapper[4799]: I1124 08:44:53.640796 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" path="/var/lib/kubelet/pods/331f65db-6ad5-4d2e-8af9-b0865ba609a7/volumes" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.165236 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7"] Nov 24 08:45:00 crc kubenswrapper[4799]: E1124 08:45:00.166343 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" containerName="registry-server" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.166364 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" containerName="registry-server" Nov 24 08:45:00 crc kubenswrapper[4799]: E1124 08:45:00.166387 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" containerName="extract-content" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.166395 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" containerName="extract-content" Nov 24 08:45:00 crc kubenswrapper[4799]: E1124 08:45:00.166412 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" containerName="extract-utilities" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.166419 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" containerName="extract-utilities" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.166683 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="331f65db-6ad5-4d2e-8af9-b0865ba609a7" containerName="registry-server" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.167899 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.176999 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.177156 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.182212 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7"] Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.336968 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-secret-volume\") pod \"collect-profiles-29399565-bs4v7\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.337205 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzxls\" (UniqueName: \"kubernetes.io/projected/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-kube-api-access-qzxls\") pod \"collect-profiles-29399565-bs4v7\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.337338 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-config-volume\") pod \"collect-profiles-29399565-bs4v7\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.438908 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-secret-volume\") pod \"collect-profiles-29399565-bs4v7\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.439097 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzxls\" (UniqueName: \"kubernetes.io/projected/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-kube-api-access-qzxls\") pod \"collect-profiles-29399565-bs4v7\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.439557 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-config-volume\") pod \"collect-profiles-29399565-bs4v7\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.440340 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-config-volume\") pod \"collect-profiles-29399565-bs4v7\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.454087 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-secret-volume\") pod \"collect-profiles-29399565-bs4v7\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.456861 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzxls\" (UniqueName: \"kubernetes.io/projected/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-kube-api-access-qzxls\") pod \"collect-profiles-29399565-bs4v7\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.490933 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:00 crc kubenswrapper[4799]: I1124 08:45:00.946689 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7"] Nov 24 08:45:01 crc kubenswrapper[4799]: I1124 08:45:01.349581 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" event={"ID":"5971752e-f0b4-4ece-b52f-e8a004ea0ff6","Type":"ContainerStarted","Data":"c19b58ad3cf69e86484938a1bcbb6e461d41cf31fef782a4dfa0dbdd0af08dce"} Nov 24 08:45:01 crc kubenswrapper[4799]: I1124 08:45:01.350001 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" event={"ID":"5971752e-f0b4-4ece-b52f-e8a004ea0ff6","Type":"ContainerStarted","Data":"4adc2466fd0bd0e54330316f87c0f1fdeefc1dc174621e81f8e5d21b4f12d081"} Nov 24 08:45:01 crc kubenswrapper[4799]: I1124 08:45:01.381902 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" podStartSLOduration=1.381882962 podStartE2EDuration="1.381882962s" podCreationTimestamp="2025-11-24 08:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:45:01.371041614 +0000 UTC m=+7047.027024088" watchObservedRunningTime="2025-11-24 08:45:01.381882962 +0000 UTC m=+7047.037865436" Nov 24 08:45:02 crc kubenswrapper[4799]: I1124 08:45:02.360288 4799 generic.go:334] "Generic (PLEG): container finished" podID="5971752e-f0b4-4ece-b52f-e8a004ea0ff6" containerID="c19b58ad3cf69e86484938a1bcbb6e461d41cf31fef782a4dfa0dbdd0af08dce" exitCode=0 Nov 24 08:45:02 crc kubenswrapper[4799]: I1124 08:45:02.360340 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" event={"ID":"5971752e-f0b4-4ece-b52f-e8a004ea0ff6","Type":"ContainerDied","Data":"c19b58ad3cf69e86484938a1bcbb6e461d41cf31fef782a4dfa0dbdd0af08dce"} Nov 24 08:45:02 crc kubenswrapper[4799]: I1124 08:45:02.628598 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:45:02 crc kubenswrapper[4799]: E1124 08:45:02.628913 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:45:03 crc kubenswrapper[4799]: I1124 08:45:03.779315 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:03 crc kubenswrapper[4799]: I1124 08:45:03.907947 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzxls\" (UniqueName: \"kubernetes.io/projected/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-kube-api-access-qzxls\") pod \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " Nov 24 08:45:03 crc kubenswrapper[4799]: I1124 08:45:03.908042 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-config-volume\") pod \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " Nov 24 08:45:03 crc kubenswrapper[4799]: I1124 08:45:03.908144 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-secret-volume\") pod \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\" (UID: \"5971752e-f0b4-4ece-b52f-e8a004ea0ff6\") " Nov 24 08:45:03 crc kubenswrapper[4799]: I1124 08:45:03.908932 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-config-volume" (OuterVolumeSpecName: "config-volume") pod "5971752e-f0b4-4ece-b52f-e8a004ea0ff6" (UID: "5971752e-f0b4-4ece-b52f-e8a004ea0ff6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:45:03 crc kubenswrapper[4799]: I1124 08:45:03.913430 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-kube-api-access-qzxls" (OuterVolumeSpecName: "kube-api-access-qzxls") pod "5971752e-f0b4-4ece-b52f-e8a004ea0ff6" (UID: "5971752e-f0b4-4ece-b52f-e8a004ea0ff6"). InnerVolumeSpecName "kube-api-access-qzxls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:45:03 crc kubenswrapper[4799]: I1124 08:45:03.919136 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5971752e-f0b4-4ece-b52f-e8a004ea0ff6" (UID: "5971752e-f0b4-4ece-b52f-e8a004ea0ff6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:45:04 crc kubenswrapper[4799]: I1124 08:45:04.010618 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzxls\" (UniqueName: \"kubernetes.io/projected/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-kube-api-access-qzxls\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:04 crc kubenswrapper[4799]: I1124 08:45:04.010667 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:04 crc kubenswrapper[4799]: I1124 08:45:04.010679 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5971752e-f0b4-4ece-b52f-e8a004ea0ff6-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:04 crc kubenswrapper[4799]: I1124 08:45:04.382332 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" event={"ID":"5971752e-f0b4-4ece-b52f-e8a004ea0ff6","Type":"ContainerDied","Data":"4adc2466fd0bd0e54330316f87c0f1fdeefc1dc174621e81f8e5d21b4f12d081"} Nov 24 08:45:04 crc kubenswrapper[4799]: I1124 08:45:04.382744 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4adc2466fd0bd0e54330316f87c0f1fdeefc1dc174621e81f8e5d21b4f12d081" Nov 24 08:45:04 crc kubenswrapper[4799]: I1124 08:45:04.382458 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7" Nov 24 08:45:04 crc kubenswrapper[4799]: I1124 08:45:04.861105 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc"] Nov 24 08:45:04 crc kubenswrapper[4799]: I1124 08:45:04.870802 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399520-p4wbc"] Nov 24 08:45:05 crc kubenswrapper[4799]: I1124 08:45:05.643634 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e1170b5-6a86-479d-9f29-dcffda7f5ba7" path="/var/lib/kubelet/pods/9e1170b5-6a86-479d-9f29-dcffda7f5ba7/volumes" Nov 24 08:45:11 crc kubenswrapper[4799]: I1124 08:45:11.507063 4799 scope.go:117] "RemoveContainer" containerID="8b11a0a33522def7c70ba09d3f52e193848422479019be5c27d2278e307c7b3d" Nov 24 08:45:14 crc kubenswrapper[4799]: I1124 08:45:14.628714 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:45:14 crc kubenswrapper[4799]: E1124 08:45:14.629593 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:45:26 crc kubenswrapper[4799]: I1124 08:45:26.629524 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:45:26 crc kubenswrapper[4799]: E1124 08:45:26.630249 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:45:40 crc kubenswrapper[4799]: I1124 08:45:40.628478 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:45:40 crc kubenswrapper[4799]: E1124 08:45:40.629257 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:45:51 crc kubenswrapper[4799]: I1124 08:45:51.629664 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:45:51 crc kubenswrapper[4799]: E1124 08:45:51.630581 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:46:02 crc kubenswrapper[4799]: I1124 08:46:02.628706 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:46:02 crc kubenswrapper[4799]: E1124 08:46:02.629498 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:46:15 crc kubenswrapper[4799]: I1124 08:46:15.642468 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:46:15 crc kubenswrapper[4799]: E1124 08:46:15.643914 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:46:27 crc kubenswrapper[4799]: I1124 08:46:27.628286 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:46:27 crc kubenswrapper[4799]: E1124 08:46:27.629263 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:46:40 crc kubenswrapper[4799]: I1124 08:46:40.628951 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:46:40 crc kubenswrapper[4799]: E1124 08:46:40.629997 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:46:51 crc kubenswrapper[4799]: I1124 08:46:51.744097 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rztnn"] Nov 24 08:46:51 crc kubenswrapper[4799]: E1124 08:46:51.746364 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5971752e-f0b4-4ece-b52f-e8a004ea0ff6" containerName="collect-profiles" Nov 24 08:46:51 crc kubenswrapper[4799]: I1124 08:46:51.746384 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5971752e-f0b4-4ece-b52f-e8a004ea0ff6" containerName="collect-profiles" Nov 24 08:46:51 crc kubenswrapper[4799]: I1124 08:46:51.746681 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5971752e-f0b4-4ece-b52f-e8a004ea0ff6" containerName="collect-profiles" Nov 24 08:46:51 crc kubenswrapper[4799]: I1124 08:46:51.748879 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:46:51 crc kubenswrapper[4799]: I1124 08:46:51.754915 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rztnn"] Nov 24 08:46:51 crc kubenswrapper[4799]: I1124 08:46:51.899293 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-catalog-content\") pod \"redhat-marketplace-rztnn\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:46:51 crc kubenswrapper[4799]: I1124 08:46:51.899399 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js697\" (UniqueName: \"kubernetes.io/projected/5661dd80-588d-4ab9-8360-44b2e50871f7-kube-api-access-js697\") pod \"redhat-marketplace-rztnn\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:46:51 crc kubenswrapper[4799]: I1124 08:46:51.899448 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-utilities\") pod \"redhat-marketplace-rztnn\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:46:52 crc kubenswrapper[4799]: I1124 08:46:52.001764 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-catalog-content\") pod \"redhat-marketplace-rztnn\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:46:52 crc kubenswrapper[4799]: I1124 08:46:52.001918 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js697\" (UniqueName: \"kubernetes.io/projected/5661dd80-588d-4ab9-8360-44b2e50871f7-kube-api-access-js697\") pod \"redhat-marketplace-rztnn\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:46:52 crc kubenswrapper[4799]: I1124 08:46:52.001997 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-utilities\") pod \"redhat-marketplace-rztnn\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:46:52 crc kubenswrapper[4799]: I1124 08:46:52.002315 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-catalog-content\") pod \"redhat-marketplace-rztnn\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:46:52 crc kubenswrapper[4799]: I1124 08:46:52.002562 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-utilities\") pod \"redhat-marketplace-rztnn\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:46:52 crc kubenswrapper[4799]: I1124 08:46:52.021070 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js697\" (UniqueName: \"kubernetes.io/projected/5661dd80-588d-4ab9-8360-44b2e50871f7-kube-api-access-js697\") pod \"redhat-marketplace-rztnn\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:46:52 crc kubenswrapper[4799]: I1124 08:46:52.082716 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:46:52 crc kubenswrapper[4799]: I1124 08:46:52.573446 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rztnn"] Nov 24 08:46:52 crc kubenswrapper[4799]: I1124 08:46:52.628432 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:46:52 crc kubenswrapper[4799]: E1124 08:46:52.628709 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:46:53 crc kubenswrapper[4799]: I1124 08:46:53.464608 4799 generic.go:334] "Generic (PLEG): container finished" podID="5661dd80-588d-4ab9-8360-44b2e50871f7" containerID="78f195e081fa1979f528ef6ed4309d145df12ccefcc31a1c6af94f81b5058c18" exitCode=0 Nov 24 08:46:53 crc kubenswrapper[4799]: I1124 08:46:53.464662 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztnn" event={"ID":"5661dd80-588d-4ab9-8360-44b2e50871f7","Type":"ContainerDied","Data":"78f195e081fa1979f528ef6ed4309d145df12ccefcc31a1c6af94f81b5058c18"} Nov 24 08:46:53 crc kubenswrapper[4799]: I1124 08:46:53.464919 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztnn" event={"ID":"5661dd80-588d-4ab9-8360-44b2e50871f7","Type":"ContainerStarted","Data":"a4df2ba7eb5cc37a2f1c4b028c91b387afddc4be18eaaadc32051003790433b8"} Nov 24 08:46:54 crc kubenswrapper[4799]: I1124 08:46:54.477209 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztnn" event={"ID":"5661dd80-588d-4ab9-8360-44b2e50871f7","Type":"ContainerStarted","Data":"29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338"} Nov 24 08:46:55 crc kubenswrapper[4799]: I1124 08:46:55.488242 4799 generic.go:334] "Generic (PLEG): container finished" podID="5661dd80-588d-4ab9-8360-44b2e50871f7" containerID="29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338" exitCode=0 Nov 24 08:46:55 crc kubenswrapper[4799]: I1124 08:46:55.488304 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztnn" event={"ID":"5661dd80-588d-4ab9-8360-44b2e50871f7","Type":"ContainerDied","Data":"29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338"} Nov 24 08:46:56 crc kubenswrapper[4799]: I1124 08:46:56.498817 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztnn" event={"ID":"5661dd80-588d-4ab9-8360-44b2e50871f7","Type":"ContainerStarted","Data":"b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75"} Nov 24 08:46:56 crc kubenswrapper[4799]: I1124 08:46:56.532444 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rztnn" podStartSLOduration=3.059252922 podStartE2EDuration="5.531838429s" podCreationTimestamp="2025-11-24 08:46:51 +0000 UTC" firstStartedPulling="2025-11-24 08:46:53.466172957 +0000 UTC m=+7159.122155431" lastFinishedPulling="2025-11-24 08:46:55.938758464 +0000 UTC m=+7161.594740938" observedRunningTime="2025-11-24 08:46:56.521962139 +0000 UTC m=+7162.177944623" watchObservedRunningTime="2025-11-24 08:46:56.531838429 +0000 UTC m=+7162.187820903" Nov 24 08:46:59 crc kubenswrapper[4799]: I1124 08:46:59.530406 4799 generic.go:334] "Generic (PLEG): container finished" podID="2ebcd170-a63a-4355-9046-17e8f42f4dc8" containerID="8ce01d3eb969032db337bb3c5e2bed09004ec025e4cf57e88b8fde35788fe264" exitCode=0 Nov 24 08:46:59 crc kubenswrapper[4799]: I1124 08:46:59.530480 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" event={"ID":"2ebcd170-a63a-4355-9046-17e8f42f4dc8","Type":"ContainerDied","Data":"8ce01d3eb969032db337bb3c5e2bed09004ec025e4cf57e88b8fde35788fe264"} Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.020485 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.203057 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-combined-ca-bundle\") pod \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.203117 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-inventory\") pod \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.203149 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnm6j\" (UniqueName: \"kubernetes.io/projected/2ebcd170-a63a-4355-9046-17e8f42f4dc8-kube-api-access-vnm6j\") pod \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.203198 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ssh-key\") pod \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.203405 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ceph\") pod \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.203468 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-secret-0\") pod \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\" (UID: \"2ebcd170-a63a-4355-9046-17e8f42f4dc8\") " Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.210263 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ebcd170-a63a-4355-9046-17e8f42f4dc8-kube-api-access-vnm6j" (OuterVolumeSpecName: "kube-api-access-vnm6j") pod "2ebcd170-a63a-4355-9046-17e8f42f4dc8" (UID: "2ebcd170-a63a-4355-9046-17e8f42f4dc8"). InnerVolumeSpecName "kube-api-access-vnm6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.210808 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ceph" (OuterVolumeSpecName: "ceph") pod "2ebcd170-a63a-4355-9046-17e8f42f4dc8" (UID: "2ebcd170-a63a-4355-9046-17e8f42f4dc8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.211631 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2ebcd170-a63a-4355-9046-17e8f42f4dc8" (UID: "2ebcd170-a63a-4355-9046-17e8f42f4dc8"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.239302 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "2ebcd170-a63a-4355-9046-17e8f42f4dc8" (UID: "2ebcd170-a63a-4355-9046-17e8f42f4dc8"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.242244 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-inventory" (OuterVolumeSpecName: "inventory") pod "2ebcd170-a63a-4355-9046-17e8f42f4dc8" (UID: "2ebcd170-a63a-4355-9046-17e8f42f4dc8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.242730 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2ebcd170-a63a-4355-9046-17e8f42f4dc8" (UID: "2ebcd170-a63a-4355-9046-17e8f42f4dc8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.306535 4799 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.306722 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.306813 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnm6j\" (UniqueName: \"kubernetes.io/projected/2ebcd170-a63a-4355-9046-17e8f42f4dc8-kube-api-access-vnm6j\") on node \"crc\" DevicePath \"\"" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.306892 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.306953 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.307007 4799 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2ebcd170-a63a-4355-9046-17e8f42f4dc8-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.552886 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" event={"ID":"2ebcd170-a63a-4355-9046-17e8f42f4dc8","Type":"ContainerDied","Data":"230a14ea43994626f614382d595480944f1b008138b9a08ec11dded4ac3728b5"} Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.552956 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="230a14ea43994626f614382d595480944f1b008138b9a08ec11dded4ac3728b5" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.553237 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-nnrmm" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.647619 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-tfzxs"] Nov 24 08:47:01 crc kubenswrapper[4799]: E1124 08:47:01.647976 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ebcd170-a63a-4355-9046-17e8f42f4dc8" containerName="libvirt-openstack-openstack-cell1" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.647993 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ebcd170-a63a-4355-9046-17e8f42f4dc8" containerName="libvirt-openstack-openstack-cell1" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.648210 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ebcd170-a63a-4355-9046-17e8f42f4dc8" containerName="libvirt-openstack-openstack-cell1" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.649024 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.652213 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.652501 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.652518 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.652572 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.653067 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.653378 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.654440 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.658083 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-tfzxs"] Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.830199 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.830258 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg4hh\" (UniqueName: \"kubernetes.io/projected/9356d471-b239-437f-8db4-f9be28f1c3dd-kube-api-access-fg4hh\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.830306 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.830541 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.830699 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.830772 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.831037 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.831110 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-inventory\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.831532 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ceph\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.831634 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.831661 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.935602 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.935668 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.935740 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.935783 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg4hh\" (UniqueName: \"kubernetes.io/projected/9356d471-b239-437f-8db4-f9be28f1c3dd-kube-api-access-fg4hh\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.935824 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.935901 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.935958 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.935999 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.936066 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.936099 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-inventory\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.936172 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ceph\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.936838 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.937166 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.940202 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-inventory\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.940234 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.940377 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.940411 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.940485 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ceph\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.941126 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.947238 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.947394 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.955866 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg4hh\" (UniqueName: \"kubernetes.io/projected/9356d471-b239-437f-8db4-f9be28f1c3dd-kube-api-access-fg4hh\") pod \"nova-cell1-openstack-openstack-cell1-tfzxs\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:01 crc kubenswrapper[4799]: I1124 08:47:01.986922 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:47:02 crc kubenswrapper[4799]: I1124 08:47:02.084582 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:47:02 crc kubenswrapper[4799]: I1124 08:47:02.087244 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:47:02 crc kubenswrapper[4799]: I1124 08:47:02.143160 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:47:02 crc kubenswrapper[4799]: I1124 08:47:02.565560 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-tfzxs"] Nov 24 08:47:02 crc kubenswrapper[4799]: I1124 08:47:02.618905 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:47:02 crc kubenswrapper[4799]: I1124 08:47:02.674994 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rztnn"] Nov 24 08:47:03 crc kubenswrapper[4799]: I1124 08:47:03.576146 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" event={"ID":"9356d471-b239-437f-8db4-f9be28f1c3dd","Type":"ContainerStarted","Data":"236da7691706f96a62d58c5f6ef153dac806eadc51f297f1e8212959c810ee8c"} Nov 24 08:47:03 crc kubenswrapper[4799]: I1124 08:47:03.576765 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" event={"ID":"9356d471-b239-437f-8db4-f9be28f1c3dd","Type":"ContainerStarted","Data":"4a681fac787992e45af43369202e9b4a768a555c0b547137be434b1fed7b50b6"} Nov 24 08:47:03 crc kubenswrapper[4799]: I1124 08:47:03.600103 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" podStartSLOduration=2.12883594 podStartE2EDuration="2.600087598s" podCreationTimestamp="2025-11-24 08:47:01 +0000 UTC" firstStartedPulling="2025-11-24 08:47:02.58171689 +0000 UTC m=+7168.237699374" lastFinishedPulling="2025-11-24 08:47:03.052968558 +0000 UTC m=+7168.708951032" observedRunningTime="2025-11-24 08:47:03.596112345 +0000 UTC m=+7169.252094839" watchObservedRunningTime="2025-11-24 08:47:03.600087598 +0000 UTC m=+7169.256070072" Nov 24 08:47:04 crc kubenswrapper[4799]: I1124 08:47:04.586344 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rztnn" podUID="5661dd80-588d-4ab9-8360-44b2e50871f7" containerName="registry-server" containerID="cri-o://b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75" gracePeriod=2 Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.595087 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.597191 4799 generic.go:334] "Generic (PLEG): container finished" podID="5661dd80-588d-4ab9-8360-44b2e50871f7" containerID="b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75" exitCode=0 Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.597238 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztnn" event={"ID":"5661dd80-588d-4ab9-8360-44b2e50871f7","Type":"ContainerDied","Data":"b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75"} Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.597271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztnn" event={"ID":"5661dd80-588d-4ab9-8360-44b2e50871f7","Type":"ContainerDied","Data":"a4df2ba7eb5cc37a2f1c4b028c91b387afddc4be18eaaadc32051003790433b8"} Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.597288 4799 scope.go:117] "RemoveContainer" containerID="b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.642627 4799 scope.go:117] "RemoveContainer" containerID="29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.644681 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:47:05 crc kubenswrapper[4799]: E1124 08:47:05.644998 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.665890 4799 scope.go:117] "RemoveContainer" containerID="78f195e081fa1979f528ef6ed4309d145df12ccefcc31a1c6af94f81b5058c18" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.719112 4799 scope.go:117] "RemoveContainer" containerID="b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75" Nov 24 08:47:05 crc kubenswrapper[4799]: E1124 08:47:05.719677 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75\": container with ID starting with b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75 not found: ID does not exist" containerID="b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.719730 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75"} err="failed to get container status \"b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75\": rpc error: code = NotFound desc = could not find container \"b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75\": container with ID starting with b1801c68f92c51888b3e051e8cbcd44c6b2f9917fa5800d32c1cf49edc32bb75 not found: ID does not exist" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.719762 4799 scope.go:117] "RemoveContainer" containerID="29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338" Nov 24 08:47:05 crc kubenswrapper[4799]: E1124 08:47:05.720278 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338\": container with ID starting with 29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338 not found: ID does not exist" containerID="29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.720311 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338"} err="failed to get container status \"29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338\": rpc error: code = NotFound desc = could not find container \"29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338\": container with ID starting with 29d6f1699c43543ac47175bc3fb4ee38a4dbf1ce64cfaca4372650edb7126338 not found: ID does not exist" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.720336 4799 scope.go:117] "RemoveContainer" containerID="78f195e081fa1979f528ef6ed4309d145df12ccefcc31a1c6af94f81b5058c18" Nov 24 08:47:05 crc kubenswrapper[4799]: E1124 08:47:05.720643 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78f195e081fa1979f528ef6ed4309d145df12ccefcc31a1c6af94f81b5058c18\": container with ID starting with 78f195e081fa1979f528ef6ed4309d145df12ccefcc31a1c6af94f81b5058c18 not found: ID does not exist" containerID="78f195e081fa1979f528ef6ed4309d145df12ccefcc31a1c6af94f81b5058c18" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.720668 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78f195e081fa1979f528ef6ed4309d145df12ccefcc31a1c6af94f81b5058c18"} err="failed to get container status \"78f195e081fa1979f528ef6ed4309d145df12ccefcc31a1c6af94f81b5058c18\": rpc error: code = NotFound desc = could not find container \"78f195e081fa1979f528ef6ed4309d145df12ccefcc31a1c6af94f81b5058c18\": container with ID starting with 78f195e081fa1979f528ef6ed4309d145df12ccefcc31a1c6af94f81b5058c18 not found: ID does not exist" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.738997 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-catalog-content\") pod \"5661dd80-588d-4ab9-8360-44b2e50871f7\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.739084 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-utilities\") pod \"5661dd80-588d-4ab9-8360-44b2e50871f7\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.739405 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js697\" (UniqueName: \"kubernetes.io/projected/5661dd80-588d-4ab9-8360-44b2e50871f7-kube-api-access-js697\") pod \"5661dd80-588d-4ab9-8360-44b2e50871f7\" (UID: \"5661dd80-588d-4ab9-8360-44b2e50871f7\") " Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.740363 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-utilities" (OuterVolumeSpecName: "utilities") pod "5661dd80-588d-4ab9-8360-44b2e50871f7" (UID: "5661dd80-588d-4ab9-8360-44b2e50871f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.746920 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5661dd80-588d-4ab9-8360-44b2e50871f7-kube-api-access-js697" (OuterVolumeSpecName: "kube-api-access-js697") pod "5661dd80-588d-4ab9-8360-44b2e50871f7" (UID: "5661dd80-588d-4ab9-8360-44b2e50871f7"). InnerVolumeSpecName "kube-api-access-js697". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.759111 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5661dd80-588d-4ab9-8360-44b2e50871f7" (UID: "5661dd80-588d-4ab9-8360-44b2e50871f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.844161 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js697\" (UniqueName: \"kubernetes.io/projected/5661dd80-588d-4ab9-8360-44b2e50871f7-kube-api-access-js697\") on node \"crc\" DevicePath \"\"" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.844592 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:47:05 crc kubenswrapper[4799]: I1124 08:47:05.844699 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5661dd80-588d-4ab9-8360-44b2e50871f7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:47:06 crc kubenswrapper[4799]: I1124 08:47:06.609410 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rztnn" Nov 24 08:47:06 crc kubenswrapper[4799]: I1124 08:47:06.645379 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rztnn"] Nov 24 08:47:06 crc kubenswrapper[4799]: I1124 08:47:06.654648 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rztnn"] Nov 24 08:47:07 crc kubenswrapper[4799]: I1124 08:47:07.642111 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5661dd80-588d-4ab9-8360-44b2e50871f7" path="/var/lib/kubelet/pods/5661dd80-588d-4ab9-8360-44b2e50871f7/volumes" Nov 24 08:47:16 crc kubenswrapper[4799]: I1124 08:47:16.628306 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:47:16 crc kubenswrapper[4799]: E1124 08:47:16.629278 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:47:31 crc kubenswrapper[4799]: I1124 08:47:31.630049 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:47:31 crc kubenswrapper[4799]: E1124 08:47:31.630883 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:47:46 crc kubenswrapper[4799]: I1124 08:47:46.628579 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:47:46 crc kubenswrapper[4799]: E1124 08:47:46.629389 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:47:57 crc kubenswrapper[4799]: I1124 08:47:57.628583 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:47:57 crc kubenswrapper[4799]: E1124 08:47:57.629381 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:48:12 crc kubenswrapper[4799]: I1124 08:48:12.629011 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:48:12 crc kubenswrapper[4799]: E1124 08:48:12.629879 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:48:27 crc kubenswrapper[4799]: I1124 08:48:27.628666 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:48:27 crc kubenswrapper[4799]: E1124 08:48:27.629544 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:48:37 crc kubenswrapper[4799]: I1124 08:48:37.827595 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9vgjn"] Nov 24 08:48:37 crc kubenswrapper[4799]: E1124 08:48:37.828827 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5661dd80-588d-4ab9-8360-44b2e50871f7" containerName="extract-content" Nov 24 08:48:37 crc kubenswrapper[4799]: I1124 08:48:37.828863 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5661dd80-588d-4ab9-8360-44b2e50871f7" containerName="extract-content" Nov 24 08:48:37 crc kubenswrapper[4799]: E1124 08:48:37.828890 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5661dd80-588d-4ab9-8360-44b2e50871f7" containerName="registry-server" Nov 24 08:48:37 crc kubenswrapper[4799]: I1124 08:48:37.828898 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5661dd80-588d-4ab9-8360-44b2e50871f7" containerName="registry-server" Nov 24 08:48:37 crc kubenswrapper[4799]: E1124 08:48:37.828935 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5661dd80-588d-4ab9-8360-44b2e50871f7" containerName="extract-utilities" Nov 24 08:48:37 crc kubenswrapper[4799]: I1124 08:48:37.828945 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5661dd80-588d-4ab9-8360-44b2e50871f7" containerName="extract-utilities" Nov 24 08:48:37 crc kubenswrapper[4799]: I1124 08:48:37.829200 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5661dd80-588d-4ab9-8360-44b2e50871f7" containerName="registry-server" Nov 24 08:48:37 crc kubenswrapper[4799]: I1124 08:48:37.830988 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:37 crc kubenswrapper[4799]: I1124 08:48:37.841498 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9vgjn"] Nov 24 08:48:37 crc kubenswrapper[4799]: I1124 08:48:37.966094 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-utilities\") pod \"community-operators-9vgjn\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:37 crc kubenswrapper[4799]: I1124 08:48:37.966211 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqpzz\" (UniqueName: \"kubernetes.io/projected/322342bb-3e98-41c8-a08e-614e60bb8539-kube-api-access-nqpzz\") pod \"community-operators-9vgjn\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:37 crc kubenswrapper[4799]: I1124 08:48:37.966273 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-catalog-content\") pod \"community-operators-9vgjn\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:38 crc kubenswrapper[4799]: I1124 08:48:38.069715 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-catalog-content\") pod \"community-operators-9vgjn\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:38 crc kubenswrapper[4799]: I1124 08:48:38.069970 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-utilities\") pod \"community-operators-9vgjn\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:38 crc kubenswrapper[4799]: I1124 08:48:38.070041 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqpzz\" (UniqueName: \"kubernetes.io/projected/322342bb-3e98-41c8-a08e-614e60bb8539-kube-api-access-nqpzz\") pod \"community-operators-9vgjn\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:38 crc kubenswrapper[4799]: I1124 08:48:38.070226 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-catalog-content\") pod \"community-operators-9vgjn\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:38 crc kubenswrapper[4799]: I1124 08:48:38.070380 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-utilities\") pod \"community-operators-9vgjn\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:38 crc kubenswrapper[4799]: I1124 08:48:38.094754 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqpzz\" (UniqueName: \"kubernetes.io/projected/322342bb-3e98-41c8-a08e-614e60bb8539-kube-api-access-nqpzz\") pod \"community-operators-9vgjn\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:38 crc kubenswrapper[4799]: I1124 08:48:38.152372 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:38 crc kubenswrapper[4799]: I1124 08:48:38.735070 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9vgjn"] Nov 24 08:48:39 crc kubenswrapper[4799]: I1124 08:48:39.564258 4799 generic.go:334] "Generic (PLEG): container finished" podID="322342bb-3e98-41c8-a08e-614e60bb8539" containerID="3cbc5274e5f2bfd22669ae2548b720d237543aef21eb5b87a86e5931e4ad113a" exitCode=0 Nov 24 08:48:39 crc kubenswrapper[4799]: I1124 08:48:39.564342 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vgjn" event={"ID":"322342bb-3e98-41c8-a08e-614e60bb8539","Type":"ContainerDied","Data":"3cbc5274e5f2bfd22669ae2548b720d237543aef21eb5b87a86e5931e4ad113a"} Nov 24 08:48:39 crc kubenswrapper[4799]: I1124 08:48:39.564593 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vgjn" event={"ID":"322342bb-3e98-41c8-a08e-614e60bb8539","Type":"ContainerStarted","Data":"525d777486d3e17777390488ee4cff1fe275b4be9f96a428b9392cb3af3aa1ea"} Nov 24 08:48:39 crc kubenswrapper[4799]: I1124 08:48:39.566596 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:48:41 crc kubenswrapper[4799]: I1124 08:48:41.585285 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vgjn" event={"ID":"322342bb-3e98-41c8-a08e-614e60bb8539","Type":"ContainerStarted","Data":"b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea"} Nov 24 08:48:42 crc kubenswrapper[4799]: I1124 08:48:42.601451 4799 generic.go:334] "Generic (PLEG): container finished" podID="322342bb-3e98-41c8-a08e-614e60bb8539" containerID="b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea" exitCode=0 Nov 24 08:48:42 crc kubenswrapper[4799]: I1124 08:48:42.601504 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vgjn" event={"ID":"322342bb-3e98-41c8-a08e-614e60bb8539","Type":"ContainerDied","Data":"b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea"} Nov 24 08:48:42 crc kubenswrapper[4799]: I1124 08:48:42.628803 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:48:42 crc kubenswrapper[4799]: E1124 08:48:42.629393 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:48:43 crc kubenswrapper[4799]: I1124 08:48:43.614941 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vgjn" event={"ID":"322342bb-3e98-41c8-a08e-614e60bb8539","Type":"ContainerStarted","Data":"35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0"} Nov 24 08:48:43 crc kubenswrapper[4799]: I1124 08:48:43.637769 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9vgjn" podStartSLOduration=3.030378559 podStartE2EDuration="6.637749158s" podCreationTimestamp="2025-11-24 08:48:37 +0000 UTC" firstStartedPulling="2025-11-24 08:48:39.566323957 +0000 UTC m=+7265.222306431" lastFinishedPulling="2025-11-24 08:48:43.173694546 +0000 UTC m=+7268.829677030" observedRunningTime="2025-11-24 08:48:43.632603882 +0000 UTC m=+7269.288586387" watchObservedRunningTime="2025-11-24 08:48:43.637749158 +0000 UTC m=+7269.293731642" Nov 24 08:48:48 crc kubenswrapper[4799]: I1124 08:48:48.153145 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:48 crc kubenswrapper[4799]: I1124 08:48:48.153543 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:48 crc kubenswrapper[4799]: I1124 08:48:48.201361 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:48 crc kubenswrapper[4799]: I1124 08:48:48.727082 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:48 crc kubenswrapper[4799]: I1124 08:48:48.780598 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9vgjn"] Nov 24 08:48:50 crc kubenswrapper[4799]: I1124 08:48:50.696706 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9vgjn" podUID="322342bb-3e98-41c8-a08e-614e60bb8539" containerName="registry-server" containerID="cri-o://35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0" gracePeriod=2 Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.176118 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.322034 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-catalog-content\") pod \"322342bb-3e98-41c8-a08e-614e60bb8539\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.322268 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqpzz\" (UniqueName: \"kubernetes.io/projected/322342bb-3e98-41c8-a08e-614e60bb8539-kube-api-access-nqpzz\") pod \"322342bb-3e98-41c8-a08e-614e60bb8539\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.322330 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-utilities\") pod \"322342bb-3e98-41c8-a08e-614e60bb8539\" (UID: \"322342bb-3e98-41c8-a08e-614e60bb8539\") " Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.323471 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-utilities" (OuterVolumeSpecName: "utilities") pod "322342bb-3e98-41c8-a08e-614e60bb8539" (UID: "322342bb-3e98-41c8-a08e-614e60bb8539"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.331052 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/322342bb-3e98-41c8-a08e-614e60bb8539-kube-api-access-nqpzz" (OuterVolumeSpecName: "kube-api-access-nqpzz") pod "322342bb-3e98-41c8-a08e-614e60bb8539" (UID: "322342bb-3e98-41c8-a08e-614e60bb8539"). InnerVolumeSpecName "kube-api-access-nqpzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.372059 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "322342bb-3e98-41c8-a08e-614e60bb8539" (UID: "322342bb-3e98-41c8-a08e-614e60bb8539"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.424612 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqpzz\" (UniqueName: \"kubernetes.io/projected/322342bb-3e98-41c8-a08e-614e60bb8539-kube-api-access-nqpzz\") on node \"crc\" DevicePath \"\"" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.424659 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.424672 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/322342bb-3e98-41c8-a08e-614e60bb8539-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.706129 4799 generic.go:334] "Generic (PLEG): container finished" podID="322342bb-3e98-41c8-a08e-614e60bb8539" containerID="35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0" exitCode=0 Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.706182 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vgjn" event={"ID":"322342bb-3e98-41c8-a08e-614e60bb8539","Type":"ContainerDied","Data":"35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0"} Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.706209 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vgjn" event={"ID":"322342bb-3e98-41c8-a08e-614e60bb8539","Type":"ContainerDied","Data":"525d777486d3e17777390488ee4cff1fe275b4be9f96a428b9392cb3af3aa1ea"} Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.706226 4799 scope.go:117] "RemoveContainer" containerID="35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.706223 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vgjn" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.731657 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9vgjn"] Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.732552 4799 scope.go:117] "RemoveContainer" containerID="b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.740241 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9vgjn"] Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.755247 4799 scope.go:117] "RemoveContainer" containerID="3cbc5274e5f2bfd22669ae2548b720d237543aef21eb5b87a86e5931e4ad113a" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.798146 4799 scope.go:117] "RemoveContainer" containerID="35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0" Nov 24 08:48:51 crc kubenswrapper[4799]: E1124 08:48:51.798766 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0\": container with ID starting with 35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0 not found: ID does not exist" containerID="35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.798814 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0"} err="failed to get container status \"35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0\": rpc error: code = NotFound desc = could not find container \"35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0\": container with ID starting with 35f6480a5372327b85dcff88b353d218256b0159634993fa970ad17f45e749c0 not found: ID does not exist" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.798856 4799 scope.go:117] "RemoveContainer" containerID="b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea" Nov 24 08:48:51 crc kubenswrapper[4799]: E1124 08:48:51.799162 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea\": container with ID starting with b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea not found: ID does not exist" containerID="b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.799197 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea"} err="failed to get container status \"b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea\": rpc error: code = NotFound desc = could not find container \"b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea\": container with ID starting with b5e9b1e75346663b0bab5a2425729c538c1d83899d1651b9042d4baa515eebea not found: ID does not exist" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.799237 4799 scope.go:117] "RemoveContainer" containerID="3cbc5274e5f2bfd22669ae2548b720d237543aef21eb5b87a86e5931e4ad113a" Nov 24 08:48:51 crc kubenswrapper[4799]: E1124 08:48:51.799504 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cbc5274e5f2bfd22669ae2548b720d237543aef21eb5b87a86e5931e4ad113a\": container with ID starting with 3cbc5274e5f2bfd22669ae2548b720d237543aef21eb5b87a86e5931e4ad113a not found: ID does not exist" containerID="3cbc5274e5f2bfd22669ae2548b720d237543aef21eb5b87a86e5931e4ad113a" Nov 24 08:48:51 crc kubenswrapper[4799]: I1124 08:48:51.799532 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cbc5274e5f2bfd22669ae2548b720d237543aef21eb5b87a86e5931e4ad113a"} err="failed to get container status \"3cbc5274e5f2bfd22669ae2548b720d237543aef21eb5b87a86e5931e4ad113a\": rpc error: code = NotFound desc = could not find container \"3cbc5274e5f2bfd22669ae2548b720d237543aef21eb5b87a86e5931e4ad113a\": container with ID starting with 3cbc5274e5f2bfd22669ae2548b720d237543aef21eb5b87a86e5931e4ad113a not found: ID does not exist" Nov 24 08:48:53 crc kubenswrapper[4799]: I1124 08:48:53.643431 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="322342bb-3e98-41c8-a08e-614e60bb8539" path="/var/lib/kubelet/pods/322342bb-3e98-41c8-a08e-614e60bb8539/volumes" Nov 24 08:48:54 crc kubenswrapper[4799]: I1124 08:48:54.629243 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:48:54 crc kubenswrapper[4799]: E1124 08:48:54.629650 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:49:05 crc kubenswrapper[4799]: I1124 08:49:05.628802 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:49:05 crc kubenswrapper[4799]: E1124 08:49:05.629573 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:49:19 crc kubenswrapper[4799]: I1124 08:49:19.630573 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:49:19 crc kubenswrapper[4799]: E1124 08:49:19.632190 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:49:31 crc kubenswrapper[4799]: I1124 08:49:31.628216 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:49:31 crc kubenswrapper[4799]: E1124 08:49:31.629067 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:49:43 crc kubenswrapper[4799]: I1124 08:49:43.815573 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:49:43 crc kubenswrapper[4799]: E1124 08:49:43.818692 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:49:58 crc kubenswrapper[4799]: I1124 08:49:58.628608 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:49:58 crc kubenswrapper[4799]: I1124 08:49:58.994379 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"b10b81361deed51c59e98517b845a73cf8efa84f826574a59cf1ba8b2a388eae"} Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.820747 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6jwpm"] Nov 24 08:50:00 crc kubenswrapper[4799]: E1124 08:50:00.821679 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322342bb-3e98-41c8-a08e-614e60bb8539" containerName="extract-content" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.821693 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="322342bb-3e98-41c8-a08e-614e60bb8539" containerName="extract-content" Nov 24 08:50:00 crc kubenswrapper[4799]: E1124 08:50:00.821721 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322342bb-3e98-41c8-a08e-614e60bb8539" containerName="extract-utilities" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.821727 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="322342bb-3e98-41c8-a08e-614e60bb8539" containerName="extract-utilities" Nov 24 08:50:00 crc kubenswrapper[4799]: E1124 08:50:00.821737 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322342bb-3e98-41c8-a08e-614e60bb8539" containerName="registry-server" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.821743 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="322342bb-3e98-41c8-a08e-614e60bb8539" containerName="registry-server" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.821959 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="322342bb-3e98-41c8-a08e-614e60bb8539" containerName="registry-server" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.823431 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.837075 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-utilities\") pod \"redhat-operators-6jwpm\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.837143 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp2mb\" (UniqueName: \"kubernetes.io/projected/b4b81a95-7902-423b-a982-a68baa73d9ff-kube-api-access-fp2mb\") pod \"redhat-operators-6jwpm\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.837494 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-catalog-content\") pod \"redhat-operators-6jwpm\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.841829 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6jwpm"] Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.940477 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-catalog-content\") pod \"redhat-operators-6jwpm\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.940633 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-utilities\") pod \"redhat-operators-6jwpm\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.940666 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp2mb\" (UniqueName: \"kubernetes.io/projected/b4b81a95-7902-423b-a982-a68baa73d9ff-kube-api-access-fp2mb\") pod \"redhat-operators-6jwpm\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.941049 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-catalog-content\") pod \"redhat-operators-6jwpm\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.941065 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-utilities\") pod \"redhat-operators-6jwpm\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:00 crc kubenswrapper[4799]: I1124 08:50:00.962639 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp2mb\" (UniqueName: \"kubernetes.io/projected/b4b81a95-7902-423b-a982-a68baa73d9ff-kube-api-access-fp2mb\") pod \"redhat-operators-6jwpm\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:01 crc kubenswrapper[4799]: I1124 08:50:01.146524 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:01 crc kubenswrapper[4799]: I1124 08:50:01.711377 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6jwpm"] Nov 24 08:50:02 crc kubenswrapper[4799]: I1124 08:50:02.022910 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jwpm" event={"ID":"b4b81a95-7902-423b-a982-a68baa73d9ff","Type":"ContainerStarted","Data":"63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094"} Nov 24 08:50:02 crc kubenswrapper[4799]: I1124 08:50:02.023274 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jwpm" event={"ID":"b4b81a95-7902-423b-a982-a68baa73d9ff","Type":"ContainerStarted","Data":"3c2bad77c48fbf1f94e459fa82e44b0390494d6ddad9f143c8ce2a63766da9b5"} Nov 24 08:50:03 crc kubenswrapper[4799]: I1124 08:50:03.033542 4799 generic.go:334] "Generic (PLEG): container finished" podID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerID="63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094" exitCode=0 Nov 24 08:50:03 crc kubenswrapper[4799]: I1124 08:50:03.033619 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jwpm" event={"ID":"b4b81a95-7902-423b-a982-a68baa73d9ff","Type":"ContainerDied","Data":"63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094"} Nov 24 08:50:04 crc kubenswrapper[4799]: I1124 08:50:04.044923 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jwpm" event={"ID":"b4b81a95-7902-423b-a982-a68baa73d9ff","Type":"ContainerStarted","Data":"2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea"} Nov 24 08:50:07 crc kubenswrapper[4799]: I1124 08:50:07.085556 4799 generic.go:334] "Generic (PLEG): container finished" podID="9356d471-b239-437f-8db4-f9be28f1c3dd" containerID="236da7691706f96a62d58c5f6ef153dac806eadc51f297f1e8212959c810ee8c" exitCode=0 Nov 24 08:50:07 crc kubenswrapper[4799]: I1124 08:50:07.085647 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" event={"ID":"9356d471-b239-437f-8db4-f9be28f1c3dd","Type":"ContainerDied","Data":"236da7691706f96a62d58c5f6ef153dac806eadc51f297f1e8212959c810ee8c"} Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.657971 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.812143 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ceph\") pod \"9356d471-b239-437f-8db4-f9be28f1c3dd\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.812439 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-combined-ca-bundle\") pod \"9356d471-b239-437f-8db4-f9be28f1c3dd\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.812483 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-1\") pod \"9356d471-b239-437f-8db4-f9be28f1c3dd\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.812506 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-inventory\") pod \"9356d471-b239-437f-8db4-f9be28f1c3dd\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.812529 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fg4hh\" (UniqueName: \"kubernetes.io/projected/9356d471-b239-437f-8db4-f9be28f1c3dd-kube-api-access-fg4hh\") pod \"9356d471-b239-437f-8db4-f9be28f1c3dd\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.812689 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-0\") pod \"9356d471-b239-437f-8db4-f9be28f1c3dd\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.812711 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-0\") pod \"9356d471-b239-437f-8db4-f9be28f1c3dd\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.812778 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-1\") pod \"9356d471-b239-437f-8db4-f9be28f1c3dd\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.812804 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-0\") pod \"9356d471-b239-437f-8db4-f9be28f1c3dd\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.812911 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-1\") pod \"9356d471-b239-437f-8db4-f9be28f1c3dd\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.812933 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ssh-key\") pod \"9356d471-b239-437f-8db4-f9be28f1c3dd\" (UID: \"9356d471-b239-437f-8db4-f9be28f1c3dd\") " Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.818633 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9356d471-b239-437f-8db4-f9be28f1c3dd-kube-api-access-fg4hh" (OuterVolumeSpecName: "kube-api-access-fg4hh") pod "9356d471-b239-437f-8db4-f9be28f1c3dd" (UID: "9356d471-b239-437f-8db4-f9be28f1c3dd"). InnerVolumeSpecName "kube-api-access-fg4hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.823157 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "9356d471-b239-437f-8db4-f9be28f1c3dd" (UID: "9356d471-b239-437f-8db4-f9be28f1c3dd"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.830733 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ceph" (OuterVolumeSpecName: "ceph") pod "9356d471-b239-437f-8db4-f9be28f1c3dd" (UID: "9356d471-b239-437f-8db4-f9be28f1c3dd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.842601 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9356d471-b239-437f-8db4-f9be28f1c3dd" (UID: "9356d471-b239-437f-8db4-f9be28f1c3dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.843113 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-inventory" (OuterVolumeSpecName: "inventory") pod "9356d471-b239-437f-8db4-f9be28f1c3dd" (UID: "9356d471-b239-437f-8db4-f9be28f1c3dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.844520 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "9356d471-b239-437f-8db4-f9be28f1c3dd" (UID: "9356d471-b239-437f-8db4-f9be28f1c3dd"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.850463 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "9356d471-b239-437f-8db4-f9be28f1c3dd" (UID: "9356d471-b239-437f-8db4-f9be28f1c3dd"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.855781 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "9356d471-b239-437f-8db4-f9be28f1c3dd" (UID: "9356d471-b239-437f-8db4-f9be28f1c3dd"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.856175 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "9356d471-b239-437f-8db4-f9be28f1c3dd" (UID: "9356d471-b239-437f-8db4-f9be28f1c3dd"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.858032 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "9356d471-b239-437f-8db4-f9be28f1c3dd" (UID: "9356d471-b239-437f-8db4-f9be28f1c3dd"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.861296 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "9356d471-b239-437f-8db4-f9be28f1c3dd" (UID: "9356d471-b239-437f-8db4-f9be28f1c3dd"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.915237 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.915275 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.915288 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.915300 4799 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.915312 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.915323 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.915333 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.915343 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.915355 4799 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.915367 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9356d471-b239-437f-8db4-f9be28f1c3dd-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:08 crc kubenswrapper[4799]: I1124 08:50:08.915378 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fg4hh\" (UniqueName: \"kubernetes.io/projected/9356d471-b239-437f-8db4-f9be28f1c3dd-kube-api-access-fg4hh\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.107989 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" event={"ID":"9356d471-b239-437f-8db4-f9be28f1c3dd","Type":"ContainerDied","Data":"4a681fac787992e45af43369202e9b4a768a555c0b547137be434b1fed7b50b6"} Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.108425 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a681fac787992e45af43369202e9b4a768a555c0b547137be434b1fed7b50b6" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.108041 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-tfzxs" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.110872 4799 generic.go:334] "Generic (PLEG): container finished" podID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerID="2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea" exitCode=0 Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.110899 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jwpm" event={"ID":"b4b81a95-7902-423b-a982-a68baa73d9ff","Type":"ContainerDied","Data":"2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea"} Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.222871 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-hvjgw"] Nov 24 08:50:09 crc kubenswrapper[4799]: E1124 08:50:09.223538 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9356d471-b239-437f-8db4-f9be28f1c3dd" containerName="nova-cell1-openstack-openstack-cell1" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.223560 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9356d471-b239-437f-8db4-f9be28f1c3dd" containerName="nova-cell1-openstack-openstack-cell1" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.223918 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9356d471-b239-437f-8db4-f9be28f1c3dd" containerName="nova-cell1-openstack-openstack-cell1" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.227523 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.230839 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.234132 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.234371 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.234550 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.234684 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.239384 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-hvjgw"] Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.239390 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.239716 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.240148 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcv75\" (UniqueName: \"kubernetes.io/projected/9459dbfb-9863-41b4-85e3-e18eff5eec82-kube-api-access-wcv75\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.240456 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-inventory\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.240631 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.241069 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.241600 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ssh-key\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.241797 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceph\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.343147 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.343217 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.343311 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcv75\" (UniqueName: \"kubernetes.io/projected/9459dbfb-9863-41b4-85e3-e18eff5eec82-kube-api-access-wcv75\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.343344 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-inventory\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.343389 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.343472 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.343496 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ssh-key\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.344051 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceph\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.349532 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ssh-key\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.349790 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-inventory\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.350228 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.350264 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceph\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.350910 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.352357 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.352415 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.361617 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcv75\" (UniqueName: \"kubernetes.io/projected/9459dbfb-9863-41b4-85e3-e18eff5eec82-kube-api-access-wcv75\") pod \"telemetry-openstack-openstack-cell1-hvjgw\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:09 crc kubenswrapper[4799]: I1124 08:50:09.570590 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:50:10 crc kubenswrapper[4799]: I1124 08:50:10.109393 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-hvjgw"] Nov 24 08:50:10 crc kubenswrapper[4799]: W1124 08:50:10.111316 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9459dbfb_9863_41b4_85e3_e18eff5eec82.slice/crio-664297afacdc82754bd882477116fea85c59296d16e3f10687e7fe6cbf462c09 WatchSource:0}: Error finding container 664297afacdc82754bd882477116fea85c59296d16e3f10687e7fe6cbf462c09: Status 404 returned error can't find the container with id 664297afacdc82754bd882477116fea85c59296d16e3f10687e7fe6cbf462c09 Nov 24 08:50:10 crc kubenswrapper[4799]: I1124 08:50:10.121284 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jwpm" event={"ID":"b4b81a95-7902-423b-a982-a68baa73d9ff","Type":"ContainerStarted","Data":"75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924"} Nov 24 08:50:10 crc kubenswrapper[4799]: I1124 08:50:10.138221 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6jwpm" podStartSLOduration=3.363173885 podStartE2EDuration="10.138201341s" podCreationTimestamp="2025-11-24 08:50:00 +0000 UTC" firstStartedPulling="2025-11-24 08:50:03.035839474 +0000 UTC m=+7348.691821968" lastFinishedPulling="2025-11-24 08:50:09.81086695 +0000 UTC m=+7355.466849424" observedRunningTime="2025-11-24 08:50:10.137647066 +0000 UTC m=+7355.793629540" watchObservedRunningTime="2025-11-24 08:50:10.138201341 +0000 UTC m=+7355.794183815" Nov 24 08:50:11 crc kubenswrapper[4799]: I1124 08:50:11.136247 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" event={"ID":"9459dbfb-9863-41b4-85e3-e18eff5eec82","Type":"ContainerStarted","Data":"663b24f59367d499f652ad01633d89e80bfe2de0f7099d1c2d7eb2a2ceb3e105"} Nov 24 08:50:11 crc kubenswrapper[4799]: I1124 08:50:11.136713 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" event={"ID":"9459dbfb-9863-41b4-85e3-e18eff5eec82","Type":"ContainerStarted","Data":"664297afacdc82754bd882477116fea85c59296d16e3f10687e7fe6cbf462c09"} Nov 24 08:50:11 crc kubenswrapper[4799]: I1124 08:50:11.147016 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:11 crc kubenswrapper[4799]: I1124 08:50:11.147074 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:11 crc kubenswrapper[4799]: I1124 08:50:11.160711 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" podStartSLOduration=1.725830802 podStartE2EDuration="2.160691336s" podCreationTimestamp="2025-11-24 08:50:09 +0000 UTC" firstStartedPulling="2025-11-24 08:50:10.114645633 +0000 UTC m=+7355.770628107" lastFinishedPulling="2025-11-24 08:50:10.549506167 +0000 UTC m=+7356.205488641" observedRunningTime="2025-11-24 08:50:11.154435038 +0000 UTC m=+7356.810417532" watchObservedRunningTime="2025-11-24 08:50:11.160691336 +0000 UTC m=+7356.816673810" Nov 24 08:50:12 crc kubenswrapper[4799]: I1124 08:50:12.198190 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6jwpm" podUID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerName="registry-server" probeResult="failure" output=< Nov 24 08:50:12 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 08:50:12 crc kubenswrapper[4799]: > Nov 24 08:50:21 crc kubenswrapper[4799]: I1124 08:50:21.190400 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:21 crc kubenswrapper[4799]: I1124 08:50:21.245721 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:21 crc kubenswrapper[4799]: I1124 08:50:21.425881 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6jwpm"] Nov 24 08:50:22 crc kubenswrapper[4799]: I1124 08:50:22.266984 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6jwpm" podUID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerName="registry-server" containerID="cri-o://75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924" gracePeriod=2 Nov 24 08:50:22 crc kubenswrapper[4799]: I1124 08:50:22.877530 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.044248 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-catalog-content\") pod \"b4b81a95-7902-423b-a982-a68baa73d9ff\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.044376 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-utilities\") pod \"b4b81a95-7902-423b-a982-a68baa73d9ff\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.044449 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp2mb\" (UniqueName: \"kubernetes.io/projected/b4b81a95-7902-423b-a982-a68baa73d9ff-kube-api-access-fp2mb\") pod \"b4b81a95-7902-423b-a982-a68baa73d9ff\" (UID: \"b4b81a95-7902-423b-a982-a68baa73d9ff\") " Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.045402 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-utilities" (OuterVolumeSpecName: "utilities") pod "b4b81a95-7902-423b-a982-a68baa73d9ff" (UID: "b4b81a95-7902-423b-a982-a68baa73d9ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.059093 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4b81a95-7902-423b-a982-a68baa73d9ff-kube-api-access-fp2mb" (OuterVolumeSpecName: "kube-api-access-fp2mb") pod "b4b81a95-7902-423b-a982-a68baa73d9ff" (UID: "b4b81a95-7902-423b-a982-a68baa73d9ff"). InnerVolumeSpecName "kube-api-access-fp2mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.138328 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4b81a95-7902-423b-a982-a68baa73d9ff" (UID: "b4b81a95-7902-423b-a982-a68baa73d9ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.147044 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.147087 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b81a95-7902-423b-a982-a68baa73d9ff-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.147102 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp2mb\" (UniqueName: \"kubernetes.io/projected/b4b81a95-7902-423b-a982-a68baa73d9ff-kube-api-access-fp2mb\") on node \"crc\" DevicePath \"\"" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.276965 4799 generic.go:334] "Generic (PLEG): container finished" podID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerID="75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924" exitCode=0 Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.277025 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jwpm" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.277054 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jwpm" event={"ID":"b4b81a95-7902-423b-a982-a68baa73d9ff","Type":"ContainerDied","Data":"75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924"} Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.277513 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jwpm" event={"ID":"b4b81a95-7902-423b-a982-a68baa73d9ff","Type":"ContainerDied","Data":"3c2bad77c48fbf1f94e459fa82e44b0390494d6ddad9f143c8ce2a63766da9b5"} Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.277539 4799 scope.go:117] "RemoveContainer" containerID="75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.304479 4799 scope.go:117] "RemoveContainer" containerID="2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.315040 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6jwpm"] Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.325189 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6jwpm"] Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.345607 4799 scope.go:117] "RemoveContainer" containerID="63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.390189 4799 scope.go:117] "RemoveContainer" containerID="75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924" Nov 24 08:50:23 crc kubenswrapper[4799]: E1124 08:50:23.390693 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924\": container with ID starting with 75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924 not found: ID does not exist" containerID="75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.390728 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924"} err="failed to get container status \"75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924\": rpc error: code = NotFound desc = could not find container \"75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924\": container with ID starting with 75b4c51258127e5eeb3a38fc68146462014ae5251a68380d42193c4f90c36924 not found: ID does not exist" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.390748 4799 scope.go:117] "RemoveContainer" containerID="2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea" Nov 24 08:50:23 crc kubenswrapper[4799]: E1124 08:50:23.391131 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea\": container with ID starting with 2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea not found: ID does not exist" containerID="2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.391163 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea"} err="failed to get container status \"2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea\": rpc error: code = NotFound desc = could not find container \"2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea\": container with ID starting with 2169b3fb4a3cfe1dea3b936e4f247bfad7675e57d1ffa05b5e8d187e76e2a8ea not found: ID does not exist" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.391182 4799 scope.go:117] "RemoveContainer" containerID="63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094" Nov 24 08:50:23 crc kubenswrapper[4799]: E1124 08:50:23.391576 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094\": container with ID starting with 63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094 not found: ID does not exist" containerID="63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.391633 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094"} err="failed to get container status \"63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094\": rpc error: code = NotFound desc = could not find container \"63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094\": container with ID starting with 63fa88904711075d7e6875e89ca6df08e6ab33fee852e5ae09b5ccfb0b474094 not found: ID does not exist" Nov 24 08:50:23 crc kubenswrapper[4799]: I1124 08:50:23.643679 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4b81a95-7902-423b-a982-a68baa73d9ff" path="/var/lib/kubelet/pods/b4b81a95-7902-423b-a982-a68baa73d9ff/volumes" Nov 24 08:52:20 crc kubenswrapper[4799]: I1124 08:52:20.400571 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:52:20 crc kubenswrapper[4799]: I1124 08:52:20.402286 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:52:50 crc kubenswrapper[4799]: I1124 08:52:50.400320 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:52:50 crc kubenswrapper[4799]: I1124 08:52:50.400836 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:53:20 crc kubenswrapper[4799]: I1124 08:53:20.400624 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:53:20 crc kubenswrapper[4799]: I1124 08:53:20.401423 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:53:20 crc kubenswrapper[4799]: I1124 08:53:20.401564 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 08:53:20 crc kubenswrapper[4799]: I1124 08:53:20.403323 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b10b81361deed51c59e98517b845a73cf8efa84f826574a59cf1ba8b2a388eae"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:53:20 crc kubenswrapper[4799]: I1124 08:53:20.403425 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://b10b81361deed51c59e98517b845a73cf8efa84f826574a59cf1ba8b2a388eae" gracePeriod=600 Nov 24 08:53:21 crc kubenswrapper[4799]: I1124 08:53:21.385612 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="b10b81361deed51c59e98517b845a73cf8efa84f826574a59cf1ba8b2a388eae" exitCode=0 Nov 24 08:53:21 crc kubenswrapper[4799]: I1124 08:53:21.385771 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"b10b81361deed51c59e98517b845a73cf8efa84f826574a59cf1ba8b2a388eae"} Nov 24 08:53:21 crc kubenswrapper[4799]: I1124 08:53:21.386633 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b"} Nov 24 08:53:21 crc kubenswrapper[4799]: I1124 08:53:21.386745 4799 scope.go:117] "RemoveContainer" containerID="93fe1ea6c35280f9e01d0184f85711ede20bdb6f77539f578d4e810c94db8953" Nov 24 08:54:00 crc kubenswrapper[4799]: I1124 08:54:00.774107 4799 generic.go:334] "Generic (PLEG): container finished" podID="9459dbfb-9863-41b4-85e3-e18eff5eec82" containerID="663b24f59367d499f652ad01633d89e80bfe2de0f7099d1c2d7eb2a2ceb3e105" exitCode=0 Nov 24 08:54:00 crc kubenswrapper[4799]: I1124 08:54:00.774204 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" event={"ID":"9459dbfb-9863-41b4-85e3-e18eff5eec82","Type":"ContainerDied","Data":"663b24f59367d499f652ad01633d89e80bfe2de0f7099d1c2d7eb2a2ceb3e105"} Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.242964 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.310181 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceph\") pod \"9459dbfb-9863-41b4-85e3-e18eff5eec82\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.310342 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-0\") pod \"9459dbfb-9863-41b4-85e3-e18eff5eec82\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.310393 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcv75\" (UniqueName: \"kubernetes.io/projected/9459dbfb-9863-41b4-85e3-e18eff5eec82-kube-api-access-wcv75\") pod \"9459dbfb-9863-41b4-85e3-e18eff5eec82\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.310423 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-telemetry-combined-ca-bundle\") pod \"9459dbfb-9863-41b4-85e3-e18eff5eec82\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.310515 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-2\") pod \"9459dbfb-9863-41b4-85e3-e18eff5eec82\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.310610 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-1\") pod \"9459dbfb-9863-41b4-85e3-e18eff5eec82\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.310649 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ssh-key\") pod \"9459dbfb-9863-41b4-85e3-e18eff5eec82\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.310765 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-inventory\") pod \"9459dbfb-9863-41b4-85e3-e18eff5eec82\" (UID: \"9459dbfb-9863-41b4-85e3-e18eff5eec82\") " Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.316942 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "9459dbfb-9863-41b4-85e3-e18eff5eec82" (UID: "9459dbfb-9863-41b4-85e3-e18eff5eec82"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.317363 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceph" (OuterVolumeSpecName: "ceph") pod "9459dbfb-9863-41b4-85e3-e18eff5eec82" (UID: "9459dbfb-9863-41b4-85e3-e18eff5eec82"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.326347 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9459dbfb-9863-41b4-85e3-e18eff5eec82-kube-api-access-wcv75" (OuterVolumeSpecName: "kube-api-access-wcv75") pod "9459dbfb-9863-41b4-85e3-e18eff5eec82" (UID: "9459dbfb-9863-41b4-85e3-e18eff5eec82"). InnerVolumeSpecName "kube-api-access-wcv75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.341187 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "9459dbfb-9863-41b4-85e3-e18eff5eec82" (UID: "9459dbfb-9863-41b4-85e3-e18eff5eec82"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.341607 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "9459dbfb-9863-41b4-85e3-e18eff5eec82" (UID: "9459dbfb-9863-41b4-85e3-e18eff5eec82"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.350456 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9459dbfb-9863-41b4-85e3-e18eff5eec82" (UID: "9459dbfb-9863-41b4-85e3-e18eff5eec82"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.354015 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "9459dbfb-9863-41b4-85e3-e18eff5eec82" (UID: "9459dbfb-9863-41b4-85e3-e18eff5eec82"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.363174 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-inventory" (OuterVolumeSpecName: "inventory") pod "9459dbfb-9863-41b4-85e3-e18eff5eec82" (UID: "9459dbfb-9863-41b4-85e3-e18eff5eec82"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.413746 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.413794 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.413805 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.413818 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.413833 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcv75\" (UniqueName: \"kubernetes.io/projected/9459dbfb-9863-41b4-85e3-e18eff5eec82-kube-api-access-wcv75\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.413860 4799 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.413872 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.413883 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9459dbfb-9863-41b4-85e3-e18eff5eec82-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.798039 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" event={"ID":"9459dbfb-9863-41b4-85e3-e18eff5eec82","Type":"ContainerDied","Data":"664297afacdc82754bd882477116fea85c59296d16e3f10687e7fe6cbf462c09"} Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.798091 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="664297afacdc82754bd882477116fea85c59296d16e3f10687e7fe6cbf462c09" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.798161 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-hvjgw" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.903728 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-svg68"] Nov 24 08:54:02 crc kubenswrapper[4799]: E1124 08:54:02.904340 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9459dbfb-9863-41b4-85e3-e18eff5eec82" containerName="telemetry-openstack-openstack-cell1" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.904369 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9459dbfb-9863-41b4-85e3-e18eff5eec82" containerName="telemetry-openstack-openstack-cell1" Nov 24 08:54:02 crc kubenswrapper[4799]: E1124 08:54:02.904429 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerName="extract-utilities" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.904439 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerName="extract-utilities" Nov 24 08:54:02 crc kubenswrapper[4799]: E1124 08:54:02.904454 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerName="registry-server" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.904462 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerName="registry-server" Nov 24 08:54:02 crc kubenswrapper[4799]: E1124 08:54:02.904479 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerName="extract-content" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.904488 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerName="extract-content" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.904735 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9459dbfb-9863-41b4-85e3-e18eff5eec82" containerName="telemetry-openstack-openstack-cell1" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.904787 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4b81a95-7902-423b-a982-a68baa73d9ff" containerName="registry-server" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.905782 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.909381 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.909645 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.909834 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.911932 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.912063 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.916680 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-svg68"] Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.923393 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pmgd\" (UniqueName: \"kubernetes.io/projected/366926c4-79c8-47ec-be98-467421650c86-kube-api-access-7pmgd\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.923466 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.923562 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.923630 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.923711 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:02 crc kubenswrapper[4799]: I1124 08:54:02.923841 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.025729 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.025906 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.025963 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pmgd\" (UniqueName: \"kubernetes.io/projected/366926c4-79c8-47ec-be98-467421650c86-kube-api-access-7pmgd\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.026008 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.026103 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.026137 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.029650 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.030173 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.030990 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.030992 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.031494 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.046599 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pmgd\" (UniqueName: \"kubernetes.io/projected/366926c4-79c8-47ec-be98-467421650c86-kube-api-access-7pmgd\") pod \"neutron-sriov-openstack-openstack-cell1-svg68\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.230156 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.811397 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-svg68"] Nov 24 08:54:03 crc kubenswrapper[4799]: I1124 08:54:03.820663 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:54:04 crc kubenswrapper[4799]: I1124 08:54:04.830171 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" event={"ID":"366926c4-79c8-47ec-be98-467421650c86","Type":"ContainerStarted","Data":"6262ebbf54139a54012340bd0182d9137446f87947826a37c1a1337df54d1426"} Nov 24 08:54:04 crc kubenswrapper[4799]: I1124 08:54:04.830751 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" event={"ID":"366926c4-79c8-47ec-be98-467421650c86","Type":"ContainerStarted","Data":"2fbf6bb24a3d430a3bcc1cdb83d36b7604611efc9fff4305dfa477a1cfd11dd9"} Nov 24 08:54:04 crc kubenswrapper[4799]: I1124 08:54:04.850787 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" podStartSLOduration=2.172460918 podStartE2EDuration="2.850764732s" podCreationTimestamp="2025-11-24 08:54:02 +0000 UTC" firstStartedPulling="2025-11-24 08:54:03.820350214 +0000 UTC m=+7589.476332688" lastFinishedPulling="2025-11-24 08:54:04.498654028 +0000 UTC m=+7590.154636502" observedRunningTime="2025-11-24 08:54:04.845808422 +0000 UTC m=+7590.501790916" watchObservedRunningTime="2025-11-24 08:54:04.850764732 +0000 UTC m=+7590.506747196" Nov 24 08:55:20 crc kubenswrapper[4799]: I1124 08:55:20.401347 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:55:20 crc kubenswrapper[4799]: I1124 08:55:20.401996 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.568081 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mtqhb"] Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.571056 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.580918 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mtqhb"] Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.745228 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-catalog-content\") pod \"certified-operators-mtqhb\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.745964 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5xqg\" (UniqueName: \"kubernetes.io/projected/e497fa53-4f3b-4aa0-8f49-9d07220decb3-kube-api-access-k5xqg\") pod \"certified-operators-mtqhb\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.746201 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-utilities\") pod \"certified-operators-mtqhb\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.847583 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5xqg\" (UniqueName: \"kubernetes.io/projected/e497fa53-4f3b-4aa0-8f49-9d07220decb3-kube-api-access-k5xqg\") pod \"certified-operators-mtqhb\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.847693 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-utilities\") pod \"certified-operators-mtqhb\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.847724 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-catalog-content\") pod \"certified-operators-mtqhb\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.848400 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-catalog-content\") pod \"certified-operators-mtqhb\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.848595 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-utilities\") pod \"certified-operators-mtqhb\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.873806 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5xqg\" (UniqueName: \"kubernetes.io/projected/e497fa53-4f3b-4aa0-8f49-9d07220decb3-kube-api-access-k5xqg\") pod \"certified-operators-mtqhb\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:48 crc kubenswrapper[4799]: I1124 08:55:48.900461 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:49 crc kubenswrapper[4799]: I1124 08:55:49.426746 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mtqhb"] Nov 24 08:55:49 crc kubenswrapper[4799]: I1124 08:55:49.859456 4799 generic.go:334] "Generic (PLEG): container finished" podID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" containerID="8a366aeba3972632b951b6943eaf94c241b44e818d30a0e9b52a4bd3e9e950f0" exitCode=0 Nov 24 08:55:49 crc kubenswrapper[4799]: I1124 08:55:49.859556 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtqhb" event={"ID":"e497fa53-4f3b-4aa0-8f49-9d07220decb3","Type":"ContainerDied","Data":"8a366aeba3972632b951b6943eaf94c241b44e818d30a0e9b52a4bd3e9e950f0"} Nov 24 08:55:49 crc kubenswrapper[4799]: I1124 08:55:49.859893 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtqhb" event={"ID":"e497fa53-4f3b-4aa0-8f49-9d07220decb3","Type":"ContainerStarted","Data":"e53af3894fdece77a6ba1db51ac09dbed56d6804440638e82823f6a2e66ed29a"} Nov 24 08:55:50 crc kubenswrapper[4799]: I1124 08:55:50.400418 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:55:50 crc kubenswrapper[4799]: I1124 08:55:50.400874 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:55:50 crc kubenswrapper[4799]: I1124 08:55:50.873215 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtqhb" event={"ID":"e497fa53-4f3b-4aa0-8f49-9d07220decb3","Type":"ContainerStarted","Data":"945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728"} Nov 24 08:55:51 crc kubenswrapper[4799]: I1124 08:55:51.886314 4799 generic.go:334] "Generic (PLEG): container finished" podID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" containerID="945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728" exitCode=0 Nov 24 08:55:51 crc kubenswrapper[4799]: I1124 08:55:51.886392 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtqhb" event={"ID":"e497fa53-4f3b-4aa0-8f49-9d07220decb3","Type":"ContainerDied","Data":"945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728"} Nov 24 08:55:53 crc kubenswrapper[4799]: I1124 08:55:53.907423 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtqhb" event={"ID":"e497fa53-4f3b-4aa0-8f49-9d07220decb3","Type":"ContainerStarted","Data":"303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c"} Nov 24 08:55:53 crc kubenswrapper[4799]: I1124 08:55:53.931996 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mtqhb" podStartSLOduration=2.364962383 podStartE2EDuration="5.931978286s" podCreationTimestamp="2025-11-24 08:55:48 +0000 UTC" firstStartedPulling="2025-11-24 08:55:49.861485521 +0000 UTC m=+7695.517467995" lastFinishedPulling="2025-11-24 08:55:53.428501424 +0000 UTC m=+7699.084483898" observedRunningTime="2025-11-24 08:55:53.92648854 +0000 UTC m=+7699.582471014" watchObservedRunningTime="2025-11-24 08:55:53.931978286 +0000 UTC m=+7699.587960760" Nov 24 08:55:58 crc kubenswrapper[4799]: I1124 08:55:58.901628 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:58 crc kubenswrapper[4799]: I1124 08:55:58.902270 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:58 crc kubenswrapper[4799]: I1124 08:55:58.962167 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:59 crc kubenswrapper[4799]: I1124 08:55:59.015363 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:55:59 crc kubenswrapper[4799]: I1124 08:55:59.199864 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mtqhb"] Nov 24 08:56:00 crc kubenswrapper[4799]: I1124 08:56:00.980595 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mtqhb" podUID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" containerName="registry-server" containerID="cri-o://303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c" gracePeriod=2 Nov 24 08:56:01 crc kubenswrapper[4799]: I1124 08:56:01.496625 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:56:01 crc kubenswrapper[4799]: I1124 08:56:01.610058 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-catalog-content\") pod \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " Nov 24 08:56:01 crc kubenswrapper[4799]: I1124 08:56:01.610188 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-utilities\") pod \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " Nov 24 08:56:01 crc kubenswrapper[4799]: I1124 08:56:01.610397 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5xqg\" (UniqueName: \"kubernetes.io/projected/e497fa53-4f3b-4aa0-8f49-9d07220decb3-kube-api-access-k5xqg\") pod \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\" (UID: \"e497fa53-4f3b-4aa0-8f49-9d07220decb3\") " Nov 24 08:56:01 crc kubenswrapper[4799]: I1124 08:56:01.611961 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-utilities" (OuterVolumeSpecName: "utilities") pod "e497fa53-4f3b-4aa0-8f49-9d07220decb3" (UID: "e497fa53-4f3b-4aa0-8f49-9d07220decb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:56:01 crc kubenswrapper[4799]: I1124 08:56:01.617536 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e497fa53-4f3b-4aa0-8f49-9d07220decb3-kube-api-access-k5xqg" (OuterVolumeSpecName: "kube-api-access-k5xqg") pod "e497fa53-4f3b-4aa0-8f49-9d07220decb3" (UID: "e497fa53-4f3b-4aa0-8f49-9d07220decb3"). InnerVolumeSpecName "kube-api-access-k5xqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:56:01 crc kubenswrapper[4799]: I1124 08:56:01.714404 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:01 crc kubenswrapper[4799]: I1124 08:56:01.714882 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5xqg\" (UniqueName: \"kubernetes.io/projected/e497fa53-4f3b-4aa0-8f49-9d07220decb3-kube-api-access-k5xqg\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.004146 4799 generic.go:334] "Generic (PLEG): container finished" podID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" containerID="303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c" exitCode=0 Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.004208 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtqhb" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.004201 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtqhb" event={"ID":"e497fa53-4f3b-4aa0-8f49-9d07220decb3","Type":"ContainerDied","Data":"303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c"} Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.004339 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtqhb" event={"ID":"e497fa53-4f3b-4aa0-8f49-9d07220decb3","Type":"ContainerDied","Data":"e53af3894fdece77a6ba1db51ac09dbed56d6804440638e82823f6a2e66ed29a"} Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.004371 4799 scope.go:117] "RemoveContainer" containerID="303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.023866 4799 scope.go:117] "RemoveContainer" containerID="945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.053731 4799 scope.go:117] "RemoveContainer" containerID="8a366aeba3972632b951b6943eaf94c241b44e818d30a0e9b52a4bd3e9e950f0" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.094623 4799 scope.go:117] "RemoveContainer" containerID="303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c" Nov 24 08:56:02 crc kubenswrapper[4799]: E1124 08:56:02.095225 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c\": container with ID starting with 303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c not found: ID does not exist" containerID="303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.095298 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c"} err="failed to get container status \"303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c\": rpc error: code = NotFound desc = could not find container \"303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c\": container with ID starting with 303b9856edd76f1389d22ee393fdc0cb5c75808b0214887e03ec2d7374b2893c not found: ID does not exist" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.095341 4799 scope.go:117] "RemoveContainer" containerID="945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728" Nov 24 08:56:02 crc kubenswrapper[4799]: E1124 08:56:02.095813 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728\": container with ID starting with 945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728 not found: ID does not exist" containerID="945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.095883 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728"} err="failed to get container status \"945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728\": rpc error: code = NotFound desc = could not find container \"945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728\": container with ID starting with 945c55abaeafb06edf8ff57c9fa0da283e44715cbdf9b13ba0c8451f482a6728 not found: ID does not exist" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.095916 4799 scope.go:117] "RemoveContainer" containerID="8a366aeba3972632b951b6943eaf94c241b44e818d30a0e9b52a4bd3e9e950f0" Nov 24 08:56:02 crc kubenswrapper[4799]: E1124 08:56:02.096253 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a366aeba3972632b951b6943eaf94c241b44e818d30a0e9b52a4bd3e9e950f0\": container with ID starting with 8a366aeba3972632b951b6943eaf94c241b44e818d30a0e9b52a4bd3e9e950f0 not found: ID does not exist" containerID="8a366aeba3972632b951b6943eaf94c241b44e818d30a0e9b52a4bd3e9e950f0" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.096312 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a366aeba3972632b951b6943eaf94c241b44e818d30a0e9b52a4bd3e9e950f0"} err="failed to get container status \"8a366aeba3972632b951b6943eaf94c241b44e818d30a0e9b52a4bd3e9e950f0\": rpc error: code = NotFound desc = could not find container \"8a366aeba3972632b951b6943eaf94c241b44e818d30a0e9b52a4bd3e9e950f0\": container with ID starting with 8a366aeba3972632b951b6943eaf94c241b44e818d30a0e9b52a4bd3e9e950f0 not found: ID does not exist" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.296969 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e497fa53-4f3b-4aa0-8f49-9d07220decb3" (UID: "e497fa53-4f3b-4aa0-8f49-9d07220decb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.326189 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e497fa53-4f3b-4aa0-8f49-9d07220decb3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.343410 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mtqhb"] Nov 24 08:56:02 crc kubenswrapper[4799]: I1124 08:56:02.352902 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mtqhb"] Nov 24 08:56:03 crc kubenswrapper[4799]: I1124 08:56:03.643864 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" path="/var/lib/kubelet/pods/e497fa53-4f3b-4aa0-8f49-9d07220decb3/volumes" Nov 24 08:56:20 crc kubenswrapper[4799]: I1124 08:56:20.400233 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:56:20 crc kubenswrapper[4799]: I1124 08:56:20.400737 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:56:20 crc kubenswrapper[4799]: I1124 08:56:20.400781 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 08:56:20 crc kubenswrapper[4799]: I1124 08:56:20.401560 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:56:20 crc kubenswrapper[4799]: I1124 08:56:20.401613 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" gracePeriod=600 Nov 24 08:56:20 crc kubenswrapper[4799]: E1124 08:56:20.521846 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:56:21 crc kubenswrapper[4799]: I1124 08:56:21.186628 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" exitCode=0 Nov 24 08:56:21 crc kubenswrapper[4799]: I1124 08:56:21.186679 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b"} Nov 24 08:56:21 crc kubenswrapper[4799]: I1124 08:56:21.187156 4799 scope.go:117] "RemoveContainer" containerID="b10b81361deed51c59e98517b845a73cf8efa84f826574a59cf1ba8b2a388eae" Nov 24 08:56:21 crc kubenswrapper[4799]: I1124 08:56:21.188070 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:56:21 crc kubenswrapper[4799]: E1124 08:56:21.188511 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:56:32 crc kubenswrapper[4799]: I1124 08:56:32.628883 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:56:32 crc kubenswrapper[4799]: E1124 08:56:32.629796 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:56:44 crc kubenswrapper[4799]: I1124 08:56:44.628493 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:56:44 crc kubenswrapper[4799]: E1124 08:56:44.629290 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:56:59 crc kubenswrapper[4799]: I1124 08:56:59.629187 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:56:59 crc kubenswrapper[4799]: E1124 08:56:59.630155 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.158125 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rmwzr"] Nov 24 08:57:09 crc kubenswrapper[4799]: E1124 08:57:09.159287 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" containerName="extract-content" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.159305 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" containerName="extract-content" Nov 24 08:57:09 crc kubenswrapper[4799]: E1124 08:57:09.159326 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" containerName="registry-server" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.159333 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" containerName="registry-server" Nov 24 08:57:09 crc kubenswrapper[4799]: E1124 08:57:09.159368 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" containerName="extract-utilities" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.159377 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" containerName="extract-utilities" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.159642 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e497fa53-4f3b-4aa0-8f49-9d07220decb3" containerName="registry-server" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.163151 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.167752 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmwzr"] Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.262344 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75q87\" (UniqueName: \"kubernetes.io/projected/6110e438-0376-4f73-8870-a82b8e05c5c2-kube-api-access-75q87\") pod \"redhat-marketplace-rmwzr\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.262435 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-catalog-content\") pod \"redhat-marketplace-rmwzr\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.262503 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-utilities\") pod \"redhat-marketplace-rmwzr\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.364577 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75q87\" (UniqueName: \"kubernetes.io/projected/6110e438-0376-4f73-8870-a82b8e05c5c2-kube-api-access-75q87\") pod \"redhat-marketplace-rmwzr\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.364677 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-catalog-content\") pod \"redhat-marketplace-rmwzr\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.364726 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-utilities\") pod \"redhat-marketplace-rmwzr\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.365258 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-utilities\") pod \"redhat-marketplace-rmwzr\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.365611 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-catalog-content\") pod \"redhat-marketplace-rmwzr\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.387953 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75q87\" (UniqueName: \"kubernetes.io/projected/6110e438-0376-4f73-8870-a82b8e05c5c2-kube-api-access-75q87\") pod \"redhat-marketplace-rmwzr\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.488504 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:09 crc kubenswrapper[4799]: I1124 08:57:09.987601 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmwzr"] Nov 24 08:57:09 crc kubenswrapper[4799]: W1124 08:57:09.996588 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6110e438_0376_4f73_8870_a82b8e05c5c2.slice/crio-243667cf206fb2d935ef44e980ece838300ea0bde212c820ee683fc17b62274e WatchSource:0}: Error finding container 243667cf206fb2d935ef44e980ece838300ea0bde212c820ee683fc17b62274e: Status 404 returned error can't find the container with id 243667cf206fb2d935ef44e980ece838300ea0bde212c820ee683fc17b62274e Nov 24 08:57:10 crc kubenswrapper[4799]: I1124 08:57:10.708504 4799 generic.go:334] "Generic (PLEG): container finished" podID="6110e438-0376-4f73-8870-a82b8e05c5c2" containerID="706456db42b4173ceb3115bc2118654fc4a6688e9063f474dbd8977592f8a13e" exitCode=0 Nov 24 08:57:10 crc kubenswrapper[4799]: I1124 08:57:10.708681 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmwzr" event={"ID":"6110e438-0376-4f73-8870-a82b8e05c5c2","Type":"ContainerDied","Data":"706456db42b4173ceb3115bc2118654fc4a6688e9063f474dbd8977592f8a13e"} Nov 24 08:57:10 crc kubenswrapper[4799]: I1124 08:57:10.708785 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmwzr" event={"ID":"6110e438-0376-4f73-8870-a82b8e05c5c2","Type":"ContainerStarted","Data":"243667cf206fb2d935ef44e980ece838300ea0bde212c820ee683fc17b62274e"} Nov 24 08:57:11 crc kubenswrapper[4799]: I1124 08:57:11.629053 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:57:11 crc kubenswrapper[4799]: E1124 08:57:11.629731 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:57:11 crc kubenswrapper[4799]: I1124 08:57:11.722959 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmwzr" event={"ID":"6110e438-0376-4f73-8870-a82b8e05c5c2","Type":"ContainerStarted","Data":"10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d"} Nov 24 08:57:12 crc kubenswrapper[4799]: I1124 08:57:12.737553 4799 generic.go:334] "Generic (PLEG): container finished" podID="6110e438-0376-4f73-8870-a82b8e05c5c2" containerID="10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d" exitCode=0 Nov 24 08:57:12 crc kubenswrapper[4799]: I1124 08:57:12.737604 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmwzr" event={"ID":"6110e438-0376-4f73-8870-a82b8e05c5c2","Type":"ContainerDied","Data":"10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d"} Nov 24 08:57:13 crc kubenswrapper[4799]: I1124 08:57:13.748995 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmwzr" event={"ID":"6110e438-0376-4f73-8870-a82b8e05c5c2","Type":"ContainerStarted","Data":"ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3"} Nov 24 08:57:13 crc kubenswrapper[4799]: I1124 08:57:13.780063 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rmwzr" podStartSLOduration=2.37305285 podStartE2EDuration="4.780045634s" podCreationTimestamp="2025-11-24 08:57:09 +0000 UTC" firstStartedPulling="2025-11-24 08:57:10.710714549 +0000 UTC m=+7776.366697023" lastFinishedPulling="2025-11-24 08:57:13.117707343 +0000 UTC m=+7778.773689807" observedRunningTime="2025-11-24 08:57:13.775487285 +0000 UTC m=+7779.431469779" watchObservedRunningTime="2025-11-24 08:57:13.780045634 +0000 UTC m=+7779.436028108" Nov 24 08:57:19 crc kubenswrapper[4799]: I1124 08:57:19.489563 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:19 crc kubenswrapper[4799]: I1124 08:57:19.491058 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:19 crc kubenswrapper[4799]: I1124 08:57:19.545939 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:19 crc kubenswrapper[4799]: I1124 08:57:19.850951 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:19 crc kubenswrapper[4799]: I1124 08:57:19.907761 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmwzr"] Nov 24 08:57:21 crc kubenswrapper[4799]: I1124 08:57:21.819748 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rmwzr" podUID="6110e438-0376-4f73-8870-a82b8e05c5c2" containerName="registry-server" containerID="cri-o://ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3" gracePeriod=2 Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.357074 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.453598 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-utilities\") pod \"6110e438-0376-4f73-8870-a82b8e05c5c2\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.453885 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75q87\" (UniqueName: \"kubernetes.io/projected/6110e438-0376-4f73-8870-a82b8e05c5c2-kube-api-access-75q87\") pod \"6110e438-0376-4f73-8870-a82b8e05c5c2\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.453982 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-catalog-content\") pod \"6110e438-0376-4f73-8870-a82b8e05c5c2\" (UID: \"6110e438-0376-4f73-8870-a82b8e05c5c2\") " Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.454552 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-utilities" (OuterVolumeSpecName: "utilities") pod "6110e438-0376-4f73-8870-a82b8e05c5c2" (UID: "6110e438-0376-4f73-8870-a82b8e05c5c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.454890 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.459825 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6110e438-0376-4f73-8870-a82b8e05c5c2-kube-api-access-75q87" (OuterVolumeSpecName: "kube-api-access-75q87") pod "6110e438-0376-4f73-8870-a82b8e05c5c2" (UID: "6110e438-0376-4f73-8870-a82b8e05c5c2"). InnerVolumeSpecName "kube-api-access-75q87". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.481535 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6110e438-0376-4f73-8870-a82b8e05c5c2" (UID: "6110e438-0376-4f73-8870-a82b8e05c5c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.556531 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75q87\" (UniqueName: \"kubernetes.io/projected/6110e438-0376-4f73-8870-a82b8e05c5c2-kube-api-access-75q87\") on node \"crc\" DevicePath \"\"" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.556589 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6110e438-0376-4f73-8870-a82b8e05c5c2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.834865 4799 generic.go:334] "Generic (PLEG): container finished" podID="6110e438-0376-4f73-8870-a82b8e05c5c2" containerID="ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3" exitCode=0 Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.834912 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rmwzr" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.834931 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmwzr" event={"ID":"6110e438-0376-4f73-8870-a82b8e05c5c2","Type":"ContainerDied","Data":"ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3"} Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.835213 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmwzr" event={"ID":"6110e438-0376-4f73-8870-a82b8e05c5c2","Type":"ContainerDied","Data":"243667cf206fb2d935ef44e980ece838300ea0bde212c820ee683fc17b62274e"} Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.835230 4799 scope.go:117] "RemoveContainer" containerID="ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.869192 4799 scope.go:117] "RemoveContainer" containerID="10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.870227 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmwzr"] Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.879035 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmwzr"] Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.898600 4799 scope.go:117] "RemoveContainer" containerID="706456db42b4173ceb3115bc2118654fc4a6688e9063f474dbd8977592f8a13e" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.940162 4799 scope.go:117] "RemoveContainer" containerID="ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3" Nov 24 08:57:22 crc kubenswrapper[4799]: E1124 08:57:22.940768 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3\": container with ID starting with ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3 not found: ID does not exist" containerID="ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.940813 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3"} err="failed to get container status \"ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3\": rpc error: code = NotFound desc = could not find container \"ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3\": container with ID starting with ee5f248a64563f3c7064b6e18037ddc325fa0f15afbf5c197d582fb7de8036e3 not found: ID does not exist" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.940842 4799 scope.go:117] "RemoveContainer" containerID="10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d" Nov 24 08:57:22 crc kubenswrapper[4799]: E1124 08:57:22.941191 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d\": container with ID starting with 10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d not found: ID does not exist" containerID="10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.941215 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d"} err="failed to get container status \"10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d\": rpc error: code = NotFound desc = could not find container \"10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d\": container with ID starting with 10efe917ea9c948d4f4d7ef7b4711c2a7237e6f3ceb5962542260ad19e091d9d not found: ID does not exist" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.941232 4799 scope.go:117] "RemoveContainer" containerID="706456db42b4173ceb3115bc2118654fc4a6688e9063f474dbd8977592f8a13e" Nov 24 08:57:22 crc kubenswrapper[4799]: E1124 08:57:22.941539 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"706456db42b4173ceb3115bc2118654fc4a6688e9063f474dbd8977592f8a13e\": container with ID starting with 706456db42b4173ceb3115bc2118654fc4a6688e9063f474dbd8977592f8a13e not found: ID does not exist" containerID="706456db42b4173ceb3115bc2118654fc4a6688e9063f474dbd8977592f8a13e" Nov 24 08:57:22 crc kubenswrapper[4799]: I1124 08:57:22.941574 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"706456db42b4173ceb3115bc2118654fc4a6688e9063f474dbd8977592f8a13e"} err="failed to get container status \"706456db42b4173ceb3115bc2118654fc4a6688e9063f474dbd8977592f8a13e\": rpc error: code = NotFound desc = could not find container \"706456db42b4173ceb3115bc2118654fc4a6688e9063f474dbd8977592f8a13e\": container with ID starting with 706456db42b4173ceb3115bc2118654fc4a6688e9063f474dbd8977592f8a13e not found: ID does not exist" Nov 24 08:57:23 crc kubenswrapper[4799]: I1124 08:57:23.641756 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6110e438-0376-4f73-8870-a82b8e05c5c2" path="/var/lib/kubelet/pods/6110e438-0376-4f73-8870-a82b8e05c5c2/volumes" Nov 24 08:57:25 crc kubenswrapper[4799]: I1124 08:57:25.635588 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:57:25 crc kubenswrapper[4799]: E1124 08:57:25.636179 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:57:36 crc kubenswrapper[4799]: I1124 08:57:36.628334 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:57:36 crc kubenswrapper[4799]: E1124 08:57:36.629175 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:57:47 crc kubenswrapper[4799]: I1124 08:57:47.629056 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:57:47 crc kubenswrapper[4799]: E1124 08:57:47.630013 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:57:58 crc kubenswrapper[4799]: I1124 08:57:58.628767 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:57:58 crc kubenswrapper[4799]: E1124 08:57:58.630346 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:58:09 crc kubenswrapper[4799]: I1124 08:58:09.628805 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:58:09 crc kubenswrapper[4799]: E1124 08:58:09.629691 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:58:24 crc kubenswrapper[4799]: I1124 08:58:24.629790 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:58:24 crc kubenswrapper[4799]: E1124 08:58:24.630712 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:58:36 crc kubenswrapper[4799]: I1124 08:58:36.628425 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:58:36 crc kubenswrapper[4799]: E1124 08:58:36.629340 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:58:51 crc kubenswrapper[4799]: I1124 08:58:51.629130 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:58:51 crc kubenswrapper[4799]: E1124 08:58:51.630406 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:59:02 crc kubenswrapper[4799]: I1124 08:59:02.628620 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:59:02 crc kubenswrapper[4799]: E1124 08:59:02.629210 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:59:15 crc kubenswrapper[4799]: I1124 08:59:15.657277 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:59:15 crc kubenswrapper[4799]: E1124 08:59:15.658238 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:59:16 crc kubenswrapper[4799]: I1124 08:59:16.010894 4799 generic.go:334] "Generic (PLEG): container finished" podID="366926c4-79c8-47ec-be98-467421650c86" containerID="6262ebbf54139a54012340bd0182d9137446f87947826a37c1a1337df54d1426" exitCode=0 Nov 24 08:59:16 crc kubenswrapper[4799]: I1124 08:59:16.010974 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" event={"ID":"366926c4-79c8-47ec-be98-467421650c86","Type":"ContainerDied","Data":"6262ebbf54139a54012340bd0182d9137446f87947826a37c1a1337df54d1426"} Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.525907 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.696464 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ssh-key\") pod \"366926c4-79c8-47ec-be98-467421650c86\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.696557 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-agent-neutron-config-0\") pod \"366926c4-79c8-47ec-be98-467421650c86\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.696588 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-inventory\") pod \"366926c4-79c8-47ec-be98-467421650c86\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.697218 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-combined-ca-bundle\") pod \"366926c4-79c8-47ec-be98-467421650c86\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.697414 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pmgd\" (UniqueName: \"kubernetes.io/projected/366926c4-79c8-47ec-be98-467421650c86-kube-api-access-7pmgd\") pod \"366926c4-79c8-47ec-be98-467421650c86\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.697514 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ceph\") pod \"366926c4-79c8-47ec-be98-467421650c86\" (UID: \"366926c4-79c8-47ec-be98-467421650c86\") " Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.701747 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ceph" (OuterVolumeSpecName: "ceph") pod "366926c4-79c8-47ec-be98-467421650c86" (UID: "366926c4-79c8-47ec-be98-467421650c86"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.702277 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "366926c4-79c8-47ec-be98-467421650c86" (UID: "366926c4-79c8-47ec-be98-467421650c86"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.703163 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/366926c4-79c8-47ec-be98-467421650c86-kube-api-access-7pmgd" (OuterVolumeSpecName: "kube-api-access-7pmgd") pod "366926c4-79c8-47ec-be98-467421650c86" (UID: "366926c4-79c8-47ec-be98-467421650c86"). InnerVolumeSpecName "kube-api-access-7pmgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.725943 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "366926c4-79c8-47ec-be98-467421650c86" (UID: "366926c4-79c8-47ec-be98-467421650c86"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.728354 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "366926c4-79c8-47ec-be98-467421650c86" (UID: "366926c4-79c8-47ec-be98-467421650c86"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.744597 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-inventory" (OuterVolumeSpecName: "inventory") pod "366926c4-79c8-47ec-be98-467421650c86" (UID: "366926c4-79c8-47ec-be98-467421650c86"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.801419 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.801463 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.801479 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.801492 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.801505 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366926c4-79c8-47ec-be98-467421650c86-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:59:17 crc kubenswrapper[4799]: I1124 08:59:17.801522 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pmgd\" (UniqueName: \"kubernetes.io/projected/366926c4-79c8-47ec-be98-467421650c86-kube-api-access-7pmgd\") on node \"crc\" DevicePath \"\"" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.030726 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" event={"ID":"366926c4-79c8-47ec-be98-467421650c86","Type":"ContainerDied","Data":"2fbf6bb24a3d430a3bcc1cdb83d36b7604611efc9fff4305dfa477a1cfd11dd9"} Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.031008 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fbf6bb24a3d430a3bcc1cdb83d36b7604611efc9fff4305dfa477a1cfd11dd9" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.030781 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-svg68" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.118280 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2"] Nov 24 08:59:18 crc kubenswrapper[4799]: E1124 08:59:18.119176 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6110e438-0376-4f73-8870-a82b8e05c5c2" containerName="registry-server" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.119293 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6110e438-0376-4f73-8870-a82b8e05c5c2" containerName="registry-server" Nov 24 08:59:18 crc kubenswrapper[4799]: E1124 08:59:18.119381 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6110e438-0376-4f73-8870-a82b8e05c5c2" containerName="extract-utilities" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.119448 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6110e438-0376-4f73-8870-a82b8e05c5c2" containerName="extract-utilities" Nov 24 08:59:18 crc kubenswrapper[4799]: E1124 08:59:18.119533 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366926c4-79c8-47ec-be98-467421650c86" containerName="neutron-sriov-openstack-openstack-cell1" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.119612 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="366926c4-79c8-47ec-be98-467421650c86" containerName="neutron-sriov-openstack-openstack-cell1" Nov 24 08:59:18 crc kubenswrapper[4799]: E1124 08:59:18.119700 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6110e438-0376-4f73-8870-a82b8e05c5c2" containerName="extract-content" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.119776 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6110e438-0376-4f73-8870-a82b8e05c5c2" containerName="extract-content" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.120133 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6110e438-0376-4f73-8870-a82b8e05c5c2" containerName="registry-server" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.120231 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="366926c4-79c8-47ec-be98-467421650c86" containerName="neutron-sriov-openstack-openstack-cell1" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.121197 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.123390 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.127435 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.127569 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.127710 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.128306 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2"] Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.128556 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.208403 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.208483 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.208512 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.208551 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmgp4\" (UniqueName: \"kubernetes.io/projected/4928abdf-af58-44e6-8469-b6dc46f76e93-kube-api-access-kmgp4\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.208631 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.208665 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.311345 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.311488 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.311572 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.311791 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.311881 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.312003 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmgp4\" (UniqueName: \"kubernetes.io/projected/4928abdf-af58-44e6-8469-b6dc46f76e93-kube-api-access-kmgp4\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.316780 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.317155 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.317390 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.317714 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.318403 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.333431 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmgp4\" (UniqueName: \"kubernetes.io/projected/4928abdf-af58-44e6-8469-b6dc46f76e93-kube-api-access-kmgp4\") pod \"neutron-dhcp-openstack-openstack-cell1-hnqs2\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.441360 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.964105 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2"] Nov 24 08:59:18 crc kubenswrapper[4799]: I1124 08:59:18.972200 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:59:19 crc kubenswrapper[4799]: I1124 08:59:19.043564 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" event={"ID":"4928abdf-af58-44e6-8469-b6dc46f76e93","Type":"ContainerStarted","Data":"310df3ac37305d32ce011a4f9a86e6663f67f7e4a17bc134041a4348e84a1dd4"} Nov 24 08:59:20 crc kubenswrapper[4799]: I1124 08:59:20.054769 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" event={"ID":"4928abdf-af58-44e6-8469-b6dc46f76e93","Type":"ContainerStarted","Data":"40f477b694064f9b4fc80d2f4342a22d4a0654134657ad5f6222df6a0e2ec955"} Nov 24 08:59:30 crc kubenswrapper[4799]: I1124 08:59:30.630368 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:59:30 crc kubenswrapper[4799]: E1124 08:59:30.633247 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:59:41 crc kubenswrapper[4799]: I1124 08:59:41.628465 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:59:41 crc kubenswrapper[4799]: E1124 08:59:41.629298 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 08:59:56 crc kubenswrapper[4799]: I1124 08:59:56.628997 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 08:59:56 crc kubenswrapper[4799]: E1124 08:59:56.630123 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.143661 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" podStartSLOduration=41.693368898 podStartE2EDuration="42.14363805s" podCreationTimestamp="2025-11-24 08:59:18 +0000 UTC" firstStartedPulling="2025-11-24 08:59:18.972002489 +0000 UTC m=+7904.627984963" lastFinishedPulling="2025-11-24 08:59:19.422271641 +0000 UTC m=+7905.078254115" observedRunningTime="2025-11-24 08:59:20.073904449 +0000 UTC m=+7905.729886933" watchObservedRunningTime="2025-11-24 09:00:00.14363805 +0000 UTC m=+7945.799620524" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.146459 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np"] Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.148301 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.150888 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.151024 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.160723 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np"] Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.201696 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k6pv\" (UniqueName: \"kubernetes.io/projected/c09de027-4ecb-4a7d-85c1-19c039d7803a-kube-api-access-9k6pv\") pod \"collect-profiles-29399580-qr6np\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.202296 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c09de027-4ecb-4a7d-85c1-19c039d7803a-config-volume\") pod \"collect-profiles-29399580-qr6np\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.202374 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c09de027-4ecb-4a7d-85c1-19c039d7803a-secret-volume\") pod \"collect-profiles-29399580-qr6np\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.304457 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k6pv\" (UniqueName: \"kubernetes.io/projected/c09de027-4ecb-4a7d-85c1-19c039d7803a-kube-api-access-9k6pv\") pod \"collect-profiles-29399580-qr6np\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.304609 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c09de027-4ecb-4a7d-85c1-19c039d7803a-config-volume\") pod \"collect-profiles-29399580-qr6np\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.304655 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c09de027-4ecb-4a7d-85c1-19c039d7803a-secret-volume\") pod \"collect-profiles-29399580-qr6np\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.305643 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c09de027-4ecb-4a7d-85c1-19c039d7803a-config-volume\") pod \"collect-profiles-29399580-qr6np\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.311518 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c09de027-4ecb-4a7d-85c1-19c039d7803a-secret-volume\") pod \"collect-profiles-29399580-qr6np\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.319313 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k6pv\" (UniqueName: \"kubernetes.io/projected/c09de027-4ecb-4a7d-85c1-19c039d7803a-kube-api-access-9k6pv\") pod \"collect-profiles-29399580-qr6np\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.490290 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:00 crc kubenswrapper[4799]: W1124 09:00:00.931503 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc09de027_4ecb_4a7d_85c1_19c039d7803a.slice/crio-5221df21faef5aa1777337b4502f7b6483243b5213f01a705bcd4d0717bd0dcc WatchSource:0}: Error finding container 5221df21faef5aa1777337b4502f7b6483243b5213f01a705bcd4d0717bd0dcc: Status 404 returned error can't find the container with id 5221df21faef5aa1777337b4502f7b6483243b5213f01a705bcd4d0717bd0dcc Nov 24 09:00:00 crc kubenswrapper[4799]: I1124 09:00:00.932516 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np"] Nov 24 09:00:01 crc kubenswrapper[4799]: I1124 09:00:01.446280 4799 generic.go:334] "Generic (PLEG): container finished" podID="c09de027-4ecb-4a7d-85c1-19c039d7803a" containerID="cc28144969b3b7e87cfb5f665522af2489d71b1f16190420a79da06d79b7d520" exitCode=0 Nov 24 09:00:01 crc kubenswrapper[4799]: I1124 09:00:01.446445 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" event={"ID":"c09de027-4ecb-4a7d-85c1-19c039d7803a","Type":"ContainerDied","Data":"cc28144969b3b7e87cfb5f665522af2489d71b1f16190420a79da06d79b7d520"} Nov 24 09:00:01 crc kubenswrapper[4799]: I1124 09:00:01.446733 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" event={"ID":"c09de027-4ecb-4a7d-85c1-19c039d7803a","Type":"ContainerStarted","Data":"5221df21faef5aa1777337b4502f7b6483243b5213f01a705bcd4d0717bd0dcc"} Nov 24 09:00:02 crc kubenswrapper[4799]: I1124 09:00:02.826497 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:02 crc kubenswrapper[4799]: I1124 09:00:02.860047 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c09de027-4ecb-4a7d-85c1-19c039d7803a-secret-volume\") pod \"c09de027-4ecb-4a7d-85c1-19c039d7803a\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " Nov 24 09:00:02 crc kubenswrapper[4799]: I1124 09:00:02.860445 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c09de027-4ecb-4a7d-85c1-19c039d7803a-config-volume\") pod \"c09de027-4ecb-4a7d-85c1-19c039d7803a\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " Nov 24 09:00:02 crc kubenswrapper[4799]: I1124 09:00:02.860630 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k6pv\" (UniqueName: \"kubernetes.io/projected/c09de027-4ecb-4a7d-85c1-19c039d7803a-kube-api-access-9k6pv\") pod \"c09de027-4ecb-4a7d-85c1-19c039d7803a\" (UID: \"c09de027-4ecb-4a7d-85c1-19c039d7803a\") " Nov 24 09:00:02 crc kubenswrapper[4799]: I1124 09:00:02.864187 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c09de027-4ecb-4a7d-85c1-19c039d7803a-config-volume" (OuterVolumeSpecName: "config-volume") pod "c09de027-4ecb-4a7d-85c1-19c039d7803a" (UID: "c09de027-4ecb-4a7d-85c1-19c039d7803a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:00:02 crc kubenswrapper[4799]: I1124 09:00:02.867740 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c09de027-4ecb-4a7d-85c1-19c039d7803a-kube-api-access-9k6pv" (OuterVolumeSpecName: "kube-api-access-9k6pv") pod "c09de027-4ecb-4a7d-85c1-19c039d7803a" (UID: "c09de027-4ecb-4a7d-85c1-19c039d7803a"). InnerVolumeSpecName "kube-api-access-9k6pv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:00:02 crc kubenswrapper[4799]: I1124 09:00:02.890029 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c09de027-4ecb-4a7d-85c1-19c039d7803a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c09de027-4ecb-4a7d-85c1-19c039d7803a" (UID: "c09de027-4ecb-4a7d-85c1-19c039d7803a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:00:02 crc kubenswrapper[4799]: I1124 09:00:02.966379 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k6pv\" (UniqueName: \"kubernetes.io/projected/c09de027-4ecb-4a7d-85c1-19c039d7803a-kube-api-access-9k6pv\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:02 crc kubenswrapper[4799]: I1124 09:00:02.966421 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c09de027-4ecb-4a7d-85c1-19c039d7803a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:02 crc kubenswrapper[4799]: I1124 09:00:02.966433 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c09de027-4ecb-4a7d-85c1-19c039d7803a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:03 crc kubenswrapper[4799]: I1124 09:00:03.468471 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" event={"ID":"c09de027-4ecb-4a7d-85c1-19c039d7803a","Type":"ContainerDied","Data":"5221df21faef5aa1777337b4502f7b6483243b5213f01a705bcd4d0717bd0dcc"} Nov 24 09:00:03 crc kubenswrapper[4799]: I1124 09:00:03.469084 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5221df21faef5aa1777337b4502f7b6483243b5213f01a705bcd4d0717bd0dcc" Nov 24 09:00:03 crc kubenswrapper[4799]: I1124 09:00:03.468527 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qr6np" Nov 24 09:00:03 crc kubenswrapper[4799]: I1124 09:00:03.916901 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb"] Nov 24 09:00:03 crc kubenswrapper[4799]: I1124 09:00:03.925906 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399535-sd8cb"] Nov 24 09:00:05 crc kubenswrapper[4799]: I1124 09:00:05.641325 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2836c1d2-0310-411e-8895-c294f6a09c89" path="/var/lib/kubelet/pods/2836c1d2-0310-411e-8895-c294f6a09c89/volumes" Nov 24 09:00:07 crc kubenswrapper[4799]: I1124 09:00:07.629121 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 09:00:07 crc kubenswrapper[4799]: E1124 09:00:07.629881 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:00:11 crc kubenswrapper[4799]: I1124 09:00:11.975553 4799 scope.go:117] "RemoveContainer" containerID="f47d0bea02c82e54e505a7f5e5169d08633bc41c6f3cf960608787e9054103b0" Nov 24 09:00:22 crc kubenswrapper[4799]: I1124 09:00:22.628287 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 09:00:22 crc kubenswrapper[4799]: E1124 09:00:22.629215 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:00:25 crc kubenswrapper[4799]: I1124 09:00:25.846776 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c9wg9"] Nov 24 09:00:25 crc kubenswrapper[4799]: E1124 09:00:25.847998 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c09de027-4ecb-4a7d-85c1-19c039d7803a" containerName="collect-profiles" Nov 24 09:00:25 crc kubenswrapper[4799]: I1124 09:00:25.848017 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c09de027-4ecb-4a7d-85c1-19c039d7803a" containerName="collect-profiles" Nov 24 09:00:25 crc kubenswrapper[4799]: I1124 09:00:25.848287 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c09de027-4ecb-4a7d-85c1-19c039d7803a" containerName="collect-profiles" Nov 24 09:00:25 crc kubenswrapper[4799]: I1124 09:00:25.850012 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:25 crc kubenswrapper[4799]: I1124 09:00:25.857309 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c9wg9"] Nov 24 09:00:26 crc kubenswrapper[4799]: I1124 09:00:26.044569 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-catalog-content\") pod \"redhat-operators-c9wg9\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:26 crc kubenswrapper[4799]: I1124 09:00:26.044991 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-utilities\") pod \"redhat-operators-c9wg9\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:26 crc kubenswrapper[4799]: I1124 09:00:26.045085 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7b2r\" (UniqueName: \"kubernetes.io/projected/2468cd3c-6898-43bd-b115-ee4b386550f7-kube-api-access-q7b2r\") pod \"redhat-operators-c9wg9\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:26 crc kubenswrapper[4799]: I1124 09:00:26.146299 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-catalog-content\") pod \"redhat-operators-c9wg9\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:26 crc kubenswrapper[4799]: I1124 09:00:26.146646 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-utilities\") pod \"redhat-operators-c9wg9\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:26 crc kubenswrapper[4799]: I1124 09:00:26.146758 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7b2r\" (UniqueName: \"kubernetes.io/projected/2468cd3c-6898-43bd-b115-ee4b386550f7-kube-api-access-q7b2r\") pod \"redhat-operators-c9wg9\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:26 crc kubenswrapper[4799]: I1124 09:00:26.146903 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-catalog-content\") pod \"redhat-operators-c9wg9\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:26 crc kubenswrapper[4799]: I1124 09:00:26.147176 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-utilities\") pod \"redhat-operators-c9wg9\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:26 crc kubenswrapper[4799]: I1124 09:00:26.170602 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7b2r\" (UniqueName: \"kubernetes.io/projected/2468cd3c-6898-43bd-b115-ee4b386550f7-kube-api-access-q7b2r\") pod \"redhat-operators-c9wg9\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:26 crc kubenswrapper[4799]: I1124 09:00:26.184280 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:26 crc kubenswrapper[4799]: I1124 09:00:26.714916 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c9wg9"] Nov 24 09:00:27 crc kubenswrapper[4799]: I1124 09:00:27.759845 4799 generic.go:334] "Generic (PLEG): container finished" podID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerID="7878a425aa4e566d485e83b51bb6c542cf3c5d48678b5adefc79853587cf7ebc" exitCode=0 Nov 24 09:00:27 crc kubenswrapper[4799]: I1124 09:00:27.759953 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9wg9" event={"ID":"2468cd3c-6898-43bd-b115-ee4b386550f7","Type":"ContainerDied","Data":"7878a425aa4e566d485e83b51bb6c542cf3c5d48678b5adefc79853587cf7ebc"} Nov 24 09:00:27 crc kubenswrapper[4799]: I1124 09:00:27.760167 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9wg9" event={"ID":"2468cd3c-6898-43bd-b115-ee4b386550f7","Type":"ContainerStarted","Data":"35ccec04236bb0e609a28a29e78141e177432719dca5ab55e8902436e663d7bc"} Nov 24 09:00:28 crc kubenswrapper[4799]: I1124 09:00:28.772173 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9wg9" event={"ID":"2468cd3c-6898-43bd-b115-ee4b386550f7","Type":"ContainerStarted","Data":"58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a"} Nov 24 09:00:31 crc kubenswrapper[4799]: I1124 09:00:31.802986 4799 generic.go:334] "Generic (PLEG): container finished" podID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerID="58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a" exitCode=0 Nov 24 09:00:31 crc kubenswrapper[4799]: I1124 09:00:31.803101 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9wg9" event={"ID":"2468cd3c-6898-43bd-b115-ee4b386550f7","Type":"ContainerDied","Data":"58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a"} Nov 24 09:00:32 crc kubenswrapper[4799]: I1124 09:00:32.828909 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9wg9" event={"ID":"2468cd3c-6898-43bd-b115-ee4b386550f7","Type":"ContainerStarted","Data":"4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da"} Nov 24 09:00:32 crc kubenswrapper[4799]: I1124 09:00:32.848177 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c9wg9" podStartSLOduration=3.39021751 podStartE2EDuration="7.848150675s" podCreationTimestamp="2025-11-24 09:00:25 +0000 UTC" firstStartedPulling="2025-11-24 09:00:27.761959697 +0000 UTC m=+7973.417942191" lastFinishedPulling="2025-11-24 09:00:32.219892882 +0000 UTC m=+7977.875875356" observedRunningTime="2025-11-24 09:00:32.844510375 +0000 UTC m=+7978.500492849" watchObservedRunningTime="2025-11-24 09:00:32.848150675 +0000 UTC m=+7978.504133149" Nov 24 09:00:33 crc kubenswrapper[4799]: I1124 09:00:33.628915 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 09:00:33 crc kubenswrapper[4799]: E1124 09:00:33.629499 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:00:36 crc kubenswrapper[4799]: I1124 09:00:36.184886 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:36 crc kubenswrapper[4799]: I1124 09:00:36.186076 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:37 crc kubenswrapper[4799]: I1124 09:00:37.236373 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c9wg9" podUID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerName="registry-server" probeResult="failure" output=< Nov 24 09:00:37 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 09:00:37 crc kubenswrapper[4799]: > Nov 24 09:00:46 crc kubenswrapper[4799]: I1124 09:00:46.232391 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:46 crc kubenswrapper[4799]: I1124 09:00:46.297390 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:46 crc kubenswrapper[4799]: I1124 09:00:46.468200 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c9wg9"] Nov 24 09:00:46 crc kubenswrapper[4799]: I1124 09:00:46.628726 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 09:00:46 crc kubenswrapper[4799]: E1124 09:00:46.629507 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:00:48 crc kubenswrapper[4799]: I1124 09:00:48.045199 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c9wg9" podUID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerName="registry-server" containerID="cri-o://4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da" gracePeriod=2 Nov 24 09:00:48 crc kubenswrapper[4799]: I1124 09:00:48.577996 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:48 crc kubenswrapper[4799]: I1124 09:00:48.712496 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-utilities\") pod \"2468cd3c-6898-43bd-b115-ee4b386550f7\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " Nov 24 09:00:48 crc kubenswrapper[4799]: I1124 09:00:48.712698 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-catalog-content\") pod \"2468cd3c-6898-43bd-b115-ee4b386550f7\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " Nov 24 09:00:48 crc kubenswrapper[4799]: I1124 09:00:48.712823 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7b2r\" (UniqueName: \"kubernetes.io/projected/2468cd3c-6898-43bd-b115-ee4b386550f7-kube-api-access-q7b2r\") pod \"2468cd3c-6898-43bd-b115-ee4b386550f7\" (UID: \"2468cd3c-6898-43bd-b115-ee4b386550f7\") " Nov 24 09:00:48 crc kubenswrapper[4799]: I1124 09:00:48.713806 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-utilities" (OuterVolumeSpecName: "utilities") pod "2468cd3c-6898-43bd-b115-ee4b386550f7" (UID: "2468cd3c-6898-43bd-b115-ee4b386550f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:00:48 crc kubenswrapper[4799]: I1124 09:00:48.718493 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2468cd3c-6898-43bd-b115-ee4b386550f7-kube-api-access-q7b2r" (OuterVolumeSpecName: "kube-api-access-q7b2r") pod "2468cd3c-6898-43bd-b115-ee4b386550f7" (UID: "2468cd3c-6898-43bd-b115-ee4b386550f7"). InnerVolumeSpecName "kube-api-access-q7b2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:00:48 crc kubenswrapper[4799]: I1124 09:00:48.801743 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2468cd3c-6898-43bd-b115-ee4b386550f7" (UID: "2468cd3c-6898-43bd-b115-ee4b386550f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:00:48 crc kubenswrapper[4799]: I1124 09:00:48.815268 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:48 crc kubenswrapper[4799]: I1124 09:00:48.815311 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7b2r\" (UniqueName: \"kubernetes.io/projected/2468cd3c-6898-43bd-b115-ee4b386550f7-kube-api-access-q7b2r\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:48 crc kubenswrapper[4799]: I1124 09:00:48.815322 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2468cd3c-6898-43bd-b115-ee4b386550f7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.057745 4799 generic.go:334] "Generic (PLEG): container finished" podID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerID="4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da" exitCode=0 Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.057810 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9wg9" Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.057832 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9wg9" event={"ID":"2468cd3c-6898-43bd-b115-ee4b386550f7","Type":"ContainerDied","Data":"4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da"} Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.058913 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9wg9" event={"ID":"2468cd3c-6898-43bd-b115-ee4b386550f7","Type":"ContainerDied","Data":"35ccec04236bb0e609a28a29e78141e177432719dca5ab55e8902436e663d7bc"} Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.058933 4799 scope.go:117] "RemoveContainer" containerID="4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da" Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.111806 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c9wg9"] Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.112684 4799 scope.go:117] "RemoveContainer" containerID="58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a" Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.125134 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c9wg9"] Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.136939 4799 scope.go:117] "RemoveContainer" containerID="7878a425aa4e566d485e83b51bb6c542cf3c5d48678b5adefc79853587cf7ebc" Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.178077 4799 scope.go:117] "RemoveContainer" containerID="4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da" Nov 24 09:00:49 crc kubenswrapper[4799]: E1124 09:00:49.178675 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da\": container with ID starting with 4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da not found: ID does not exist" containerID="4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da" Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.178700 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da"} err="failed to get container status \"4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da\": rpc error: code = NotFound desc = could not find container \"4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da\": container with ID starting with 4464d07a19545c52c88787f68742f3794fbfe76c1061de7fb17560e9f55d98da not found: ID does not exist" Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.178722 4799 scope.go:117] "RemoveContainer" containerID="58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a" Nov 24 09:00:49 crc kubenswrapper[4799]: E1124 09:00:49.179153 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a\": container with ID starting with 58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a not found: ID does not exist" containerID="58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a" Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.179203 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a"} err="failed to get container status \"58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a\": rpc error: code = NotFound desc = could not find container \"58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a\": container with ID starting with 58d390ee91c362e07b5845ebcef77e77d88fde8ca13135067c1994b99e703a7a not found: ID does not exist" Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.179235 4799 scope.go:117] "RemoveContainer" containerID="7878a425aa4e566d485e83b51bb6c542cf3c5d48678b5adefc79853587cf7ebc" Nov 24 09:00:49 crc kubenswrapper[4799]: E1124 09:00:49.180002 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7878a425aa4e566d485e83b51bb6c542cf3c5d48678b5adefc79853587cf7ebc\": container with ID starting with 7878a425aa4e566d485e83b51bb6c542cf3c5d48678b5adefc79853587cf7ebc not found: ID does not exist" containerID="7878a425aa4e566d485e83b51bb6c542cf3c5d48678b5adefc79853587cf7ebc" Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.180022 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7878a425aa4e566d485e83b51bb6c542cf3c5d48678b5adefc79853587cf7ebc"} err="failed to get container status \"7878a425aa4e566d485e83b51bb6c542cf3c5d48678b5adefc79853587cf7ebc\": rpc error: code = NotFound desc = could not find container \"7878a425aa4e566d485e83b51bb6c542cf3c5d48678b5adefc79853587cf7ebc\": container with ID starting with 7878a425aa4e566d485e83b51bb6c542cf3c5d48678b5adefc79853587cf7ebc not found: ID does not exist" Nov 24 09:00:49 crc kubenswrapper[4799]: I1124 09:00:49.643777 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2468cd3c-6898-43bd-b115-ee4b386550f7" path="/var/lib/kubelet/pods/2468cd3c-6898-43bd-b115-ee4b386550f7/volumes" Nov 24 09:00:58 crc kubenswrapper[4799]: I1124 09:00:58.628143 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 09:00:58 crc kubenswrapper[4799]: E1124 09:00:58.628872 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.151360 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29399581-8scpq"] Nov 24 09:01:00 crc kubenswrapper[4799]: E1124 09:01:00.152229 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerName="registry-server" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.152247 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerName="registry-server" Nov 24 09:01:00 crc kubenswrapper[4799]: E1124 09:01:00.152267 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerName="extract-utilities" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.152274 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerName="extract-utilities" Nov 24 09:01:00 crc kubenswrapper[4799]: E1124 09:01:00.152291 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerName="extract-content" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.152300 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerName="extract-content" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.152578 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2468cd3c-6898-43bd-b115-ee4b386550f7" containerName="registry-server" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.153583 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.163929 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399581-8scpq"] Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.276213 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-combined-ca-bundle\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.276367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjcm2\" (UniqueName: \"kubernetes.io/projected/adc8c7bd-9458-4094-8562-65691c517956-kube-api-access-gjcm2\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.276392 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-config-data\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.276692 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-fernet-keys\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.379025 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjcm2\" (UniqueName: \"kubernetes.io/projected/adc8c7bd-9458-4094-8562-65691c517956-kube-api-access-gjcm2\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.379081 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-config-data\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.379179 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-fernet-keys\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.379220 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-combined-ca-bundle\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.387805 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-config-data\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.388010 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-combined-ca-bundle\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.397005 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-fernet-keys\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.397087 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjcm2\" (UniqueName: \"kubernetes.io/projected/adc8c7bd-9458-4094-8562-65691c517956-kube-api-access-gjcm2\") pod \"keystone-cron-29399581-8scpq\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:00 crc kubenswrapper[4799]: I1124 09:01:00.506658 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:01 crc kubenswrapper[4799]: I1124 09:01:01.069538 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399581-8scpq"] Nov 24 09:01:01 crc kubenswrapper[4799]: I1124 09:01:01.214000 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399581-8scpq" event={"ID":"adc8c7bd-9458-4094-8562-65691c517956","Type":"ContainerStarted","Data":"c22fb64b736d61fec17304b23b6b5fe9fdbccb125a12a2ee8a7acfca02573203"} Nov 24 09:01:02 crc kubenswrapper[4799]: I1124 09:01:02.226821 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399581-8scpq" event={"ID":"adc8c7bd-9458-4094-8562-65691c517956","Type":"ContainerStarted","Data":"f3c897215aaf98894973a0910fe32befccf5c3f8c31a66c4cb54c0bbe1e6b512"} Nov 24 09:01:02 crc kubenswrapper[4799]: I1124 09:01:02.244149 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29399581-8scpq" podStartSLOduration=2.244134363 podStartE2EDuration="2.244134363s" podCreationTimestamp="2025-11-24 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:01:02.241450312 +0000 UTC m=+8007.897432796" watchObservedRunningTime="2025-11-24 09:01:02.244134363 +0000 UTC m=+8007.900116837" Nov 24 09:01:09 crc kubenswrapper[4799]: I1124 09:01:09.300582 4799 generic.go:334] "Generic (PLEG): container finished" podID="adc8c7bd-9458-4094-8562-65691c517956" containerID="f3c897215aaf98894973a0910fe32befccf5c3f8c31a66c4cb54c0bbe1e6b512" exitCode=0 Nov 24 09:01:09 crc kubenswrapper[4799]: I1124 09:01:09.300683 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399581-8scpq" event={"ID":"adc8c7bd-9458-4094-8562-65691c517956","Type":"ContainerDied","Data":"f3c897215aaf98894973a0910fe32befccf5c3f8c31a66c4cb54c0bbe1e6b512"} Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.775036 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.808672 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjcm2\" (UniqueName: \"kubernetes.io/projected/adc8c7bd-9458-4094-8562-65691c517956-kube-api-access-gjcm2\") pod \"adc8c7bd-9458-4094-8562-65691c517956\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.808730 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-config-data\") pod \"adc8c7bd-9458-4094-8562-65691c517956\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.808921 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-fernet-keys\") pod \"adc8c7bd-9458-4094-8562-65691c517956\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.809638 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-combined-ca-bundle\") pod \"adc8c7bd-9458-4094-8562-65691c517956\" (UID: \"adc8c7bd-9458-4094-8562-65691c517956\") " Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.814275 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "adc8c7bd-9458-4094-8562-65691c517956" (UID: "adc8c7bd-9458-4094-8562-65691c517956"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.818265 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adc8c7bd-9458-4094-8562-65691c517956-kube-api-access-gjcm2" (OuterVolumeSpecName: "kube-api-access-gjcm2") pod "adc8c7bd-9458-4094-8562-65691c517956" (UID: "adc8c7bd-9458-4094-8562-65691c517956"). InnerVolumeSpecName "kube-api-access-gjcm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.881160 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adc8c7bd-9458-4094-8562-65691c517956" (UID: "adc8c7bd-9458-4094-8562-65691c517956"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.903499 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-config-data" (OuterVolumeSpecName: "config-data") pod "adc8c7bd-9458-4094-8562-65691c517956" (UID: "adc8c7bd-9458-4094-8562-65691c517956"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.913180 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.913216 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.913227 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjcm2\" (UniqueName: \"kubernetes.io/projected/adc8c7bd-9458-4094-8562-65691c517956-kube-api-access-gjcm2\") on node \"crc\" DevicePath \"\"" Nov 24 09:01:10 crc kubenswrapper[4799]: I1124 09:01:10.913236 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc8c7bd-9458-4094-8562-65691c517956-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:01:11 crc kubenswrapper[4799]: I1124 09:01:11.324902 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399581-8scpq" event={"ID":"adc8c7bd-9458-4094-8562-65691c517956","Type":"ContainerDied","Data":"c22fb64b736d61fec17304b23b6b5fe9fdbccb125a12a2ee8a7acfca02573203"} Nov 24 09:01:11 crc kubenswrapper[4799]: I1124 09:01:11.324956 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c22fb64b736d61fec17304b23b6b5fe9fdbccb125a12a2ee8a7acfca02573203" Nov 24 09:01:11 crc kubenswrapper[4799]: I1124 09:01:11.324968 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399581-8scpq" Nov 24 09:01:12 crc kubenswrapper[4799]: I1124 09:01:12.628945 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 09:01:12 crc kubenswrapper[4799]: E1124 09:01:12.629535 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:01:25 crc kubenswrapper[4799]: I1124 09:01:25.635054 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 09:01:26 crc kubenswrapper[4799]: I1124 09:01:26.477716 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"6c892fde2c41f62dc151886b156c9c281cd73f67e86c585ace87fe5a19ac868d"} Nov 24 09:03:03 crc kubenswrapper[4799]: I1124 09:03:03.469414 4799 generic.go:334] "Generic (PLEG): container finished" podID="4928abdf-af58-44e6-8469-b6dc46f76e93" containerID="40f477b694064f9b4fc80d2f4342a22d4a0654134657ad5f6222df6a0e2ec955" exitCode=0 Nov 24 09:03:03 crc kubenswrapper[4799]: I1124 09:03:03.470348 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" event={"ID":"4928abdf-af58-44e6-8469-b6dc46f76e93","Type":"ContainerDied","Data":"40f477b694064f9b4fc80d2f4342a22d4a0654134657ad5f6222df6a0e2ec955"} Nov 24 09:03:04 crc kubenswrapper[4799]: I1124 09:03:04.942065 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.093247 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-combined-ca-bundle\") pod \"4928abdf-af58-44e6-8469-b6dc46f76e93\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.093474 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-inventory\") pod \"4928abdf-af58-44e6-8469-b6dc46f76e93\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.093558 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmgp4\" (UniqueName: \"kubernetes.io/projected/4928abdf-af58-44e6-8469-b6dc46f76e93-kube-api-access-kmgp4\") pod \"4928abdf-af58-44e6-8469-b6dc46f76e93\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.093584 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ssh-key\") pod \"4928abdf-af58-44e6-8469-b6dc46f76e93\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.093602 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ceph\") pod \"4928abdf-af58-44e6-8469-b6dc46f76e93\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.093646 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-agent-neutron-config-0\") pod \"4928abdf-af58-44e6-8469-b6dc46f76e93\" (UID: \"4928abdf-af58-44e6-8469-b6dc46f76e93\") " Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.100430 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4928abdf-af58-44e6-8469-b6dc46f76e93-kube-api-access-kmgp4" (OuterVolumeSpecName: "kube-api-access-kmgp4") pod "4928abdf-af58-44e6-8469-b6dc46f76e93" (UID: "4928abdf-af58-44e6-8469-b6dc46f76e93"). InnerVolumeSpecName "kube-api-access-kmgp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.100430 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "4928abdf-af58-44e6-8469-b6dc46f76e93" (UID: "4928abdf-af58-44e6-8469-b6dc46f76e93"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.102034 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ceph" (OuterVolumeSpecName: "ceph") pod "4928abdf-af58-44e6-8469-b6dc46f76e93" (UID: "4928abdf-af58-44e6-8469-b6dc46f76e93"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.122396 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4928abdf-af58-44e6-8469-b6dc46f76e93" (UID: "4928abdf-af58-44e6-8469-b6dc46f76e93"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.141815 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "4928abdf-af58-44e6-8469-b6dc46f76e93" (UID: "4928abdf-af58-44e6-8469-b6dc46f76e93"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.144141 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-inventory" (OuterVolumeSpecName: "inventory") pod "4928abdf-af58-44e6-8469-b6dc46f76e93" (UID: "4928abdf-af58-44e6-8469-b6dc46f76e93"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.197265 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.197301 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.197317 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.197333 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.197348 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4928abdf-af58-44e6-8469-b6dc46f76e93-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.197362 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmgp4\" (UniqueName: \"kubernetes.io/projected/4928abdf-af58-44e6-8469-b6dc46f76e93-kube-api-access-kmgp4\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.490010 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" event={"ID":"4928abdf-af58-44e6-8469-b6dc46f76e93","Type":"ContainerDied","Data":"310df3ac37305d32ce011a4f9a86e6663f67f7e4a17bc134041a4348e84a1dd4"} Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.490052 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="310df3ac37305d32ce011a4f9a86e6663f67f7e4a17bc134041a4348e84a1dd4" Nov 24 09:03:05 crc kubenswrapper[4799]: I1124 09:03:05.490083 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hnqs2" Nov 24 09:03:27 crc kubenswrapper[4799]: I1124 09:03:27.676060 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:03:27 crc kubenswrapper[4799]: I1124 09:03:27.676832 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="2516f20b-d128-4383-9982-699005340d35" containerName="nova-cell0-conductor-conductor" containerID="cri-o://75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286" gracePeriod=30 Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.111764 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.114953 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="2ceaea61-e8ed-455e-8020-b96783d44e72" containerName="nova-cell1-conductor-conductor" containerID="cri-o://fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7" gracePeriod=30 Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.262806 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.263177 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3966b7b5-c531-4127-82d0-8d759464708a" containerName="nova-api-log" containerID="cri-o://46251192c28f8377b93fd73ac7b5fda79e3f27f7ede96780c4906f47d9f1d9e3" gracePeriod=30 Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.263316 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3966b7b5-c531-4127-82d0-8d759464708a" containerName="nova-api-api" containerID="cri-o://53b35860f548f35f4338efbdd33b9ee83820c2b24af2065c9e20cbe4f405183b" gracePeriod=30 Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.327637 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.327941 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="19134f4a-118d-4e1c-b53d-742d563576c0" containerName="nova-scheduler-scheduler" containerID="cri-o://4eca8afcb5ad58ce56463cb107cd2c212cce92b084a87e47d7e2c55e7954ea6c" gracePeriod=30 Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.345410 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.345786 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-log" containerID="cri-o://de6e817697e87c1221c2c1aa6845284ecd7b05c5c743fa80dc4ce0ed79f8fbc0" gracePeriod=30 Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.345808 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-metadata" containerID="cri-o://bae71ae1cf9e4b2ec05692bd4e70b2a8132ec02f03bac45d32f8a3fe145198a9" gracePeriod=30 Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.478883 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26"] Nov 24 09:03:28 crc kubenswrapper[4799]: E1124 09:03:28.479294 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc8c7bd-9458-4094-8562-65691c517956" containerName="keystone-cron" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.479309 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc8c7bd-9458-4094-8562-65691c517956" containerName="keystone-cron" Nov 24 09:03:28 crc kubenswrapper[4799]: E1124 09:03:28.479325 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4928abdf-af58-44e6-8469-b6dc46f76e93" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.479332 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4928abdf-af58-44e6-8469-b6dc46f76e93" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.479536 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4928abdf-af58-44e6-8469-b6dc46f76e93" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.479558 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc8c7bd-9458-4094-8562-65691c517956" containerName="keystone-cron" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.480278 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.486738 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.487140 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.487323 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.487409 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.487450 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4fkg7" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.487587 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.487608 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.500264 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26"] Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.569044 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.569144 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.569189 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.569212 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tljwl\" (UniqueName: \"kubernetes.io/projected/4ad610ec-2cdd-4eea-8d88-18b633d28225-kube-api-access-tljwl\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.569265 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.569282 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.569304 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.569337 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.569359 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.569394 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.569429 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.671630 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.671692 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.671744 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.671885 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.672075 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.672170 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.672237 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tljwl\" (UniqueName: \"kubernetes.io/projected/4ad610ec-2cdd-4eea-8d88-18b633d28225-kube-api-access-tljwl\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.672362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.672414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.672449 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.672535 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.673758 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.674584 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.690567 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.690964 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.697175 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.700195 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.712492 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.717927 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.731573 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.732189 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tljwl\" (UniqueName: \"kubernetes.io/projected/4ad610ec-2cdd-4eea-8d88-18b633d28225-kube-api-access-tljwl\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.732491 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.812451 4799 generic.go:334] "Generic (PLEG): container finished" podID="3966b7b5-c531-4127-82d0-8d759464708a" containerID="46251192c28f8377b93fd73ac7b5fda79e3f27f7ede96780c4906f47d9f1d9e3" exitCode=143 Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.812537 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3966b7b5-c531-4127-82d0-8d759464708a","Type":"ContainerDied","Data":"46251192c28f8377b93fd73ac7b5fda79e3f27f7ede96780c4906f47d9f1d9e3"} Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.838212 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.846375 4799 generic.go:334] "Generic (PLEG): container finished" podID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerID="de6e817697e87c1221c2c1aa6845284ecd7b05c5c743fa80dc4ce0ed79f8fbc0" exitCode=143 Nov 24 09:03:28 crc kubenswrapper[4799]: I1124 09:03:28.846421 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27fb95fd-a5be-4c99-9ce3-f8617695a530","Type":"ContainerDied","Data":"de6e817697e87c1221c2c1aa6845284ecd7b05c5c743fa80dc4ce0ed79f8fbc0"} Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.445070 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26"] Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.530738 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.700668 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-combined-ca-bundle\") pod \"2ceaea61-e8ed-455e-8020-b96783d44e72\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.700848 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkghd\" (UniqueName: \"kubernetes.io/projected/2ceaea61-e8ed-455e-8020-b96783d44e72-kube-api-access-vkghd\") pod \"2ceaea61-e8ed-455e-8020-b96783d44e72\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.700918 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-config-data\") pod \"2ceaea61-e8ed-455e-8020-b96783d44e72\" (UID: \"2ceaea61-e8ed-455e-8020-b96783d44e72\") " Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.712236 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ceaea61-e8ed-455e-8020-b96783d44e72-kube-api-access-vkghd" (OuterVolumeSpecName: "kube-api-access-vkghd") pod "2ceaea61-e8ed-455e-8020-b96783d44e72" (UID: "2ceaea61-e8ed-455e-8020-b96783d44e72"). InnerVolumeSpecName "kube-api-access-vkghd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.738064 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ceaea61-e8ed-455e-8020-b96783d44e72" (UID: "2ceaea61-e8ed-455e-8020-b96783d44e72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.740988 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-config-data" (OuterVolumeSpecName: "config-data") pod "2ceaea61-e8ed-455e-8020-b96783d44e72" (UID: "2ceaea61-e8ed-455e-8020-b96783d44e72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.804343 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.804380 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkghd\" (UniqueName: \"kubernetes.io/projected/2ceaea61-e8ed-455e-8020-b96783d44e72-kube-api-access-vkghd\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.804391 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ceaea61-e8ed-455e-8020-b96783d44e72-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.857247 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" event={"ID":"4ad610ec-2cdd-4eea-8d88-18b633d28225","Type":"ContainerStarted","Data":"0119e928a960cf3f43efb224d2914c4a38e80389bbc81f3237a278bcf8a0e9f3"} Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.859384 4799 generic.go:334] "Generic (PLEG): container finished" podID="2ceaea61-e8ed-455e-8020-b96783d44e72" containerID="fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7" exitCode=0 Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.859425 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.859432 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2ceaea61-e8ed-455e-8020-b96783d44e72","Type":"ContainerDied","Data":"fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7"} Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.859464 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2ceaea61-e8ed-455e-8020-b96783d44e72","Type":"ContainerDied","Data":"6e512e3e417266c72dc43025aeb8293a5530347cbce2891afb45fd39a2a266bf"} Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.859489 4799 scope.go:117] "RemoveContainer" containerID="fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.890723 4799 scope.go:117] "RemoveContainer" containerID="fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7" Nov 24 09:03:29 crc kubenswrapper[4799]: E1124 09:03:29.891230 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7\": container with ID starting with fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7 not found: ID does not exist" containerID="fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.891263 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7"} err="failed to get container status \"fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7\": rpc error: code = NotFound desc = could not find container \"fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7\": container with ID starting with fbfb4b99883dc241431c7705014497b2f0391b32a716ffeb8baff02ce02319e7 not found: ID does not exist" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.894896 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.908025 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.925124 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:03:29 crc kubenswrapper[4799]: E1124 09:03:29.925615 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ceaea61-e8ed-455e-8020-b96783d44e72" containerName="nova-cell1-conductor-conductor" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.925633 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ceaea61-e8ed-455e-8020-b96783d44e72" containerName="nova-cell1-conductor-conductor" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.925838 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ceaea61-e8ed-455e-8020-b96783d44e72" containerName="nova-cell1-conductor-conductor" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.926640 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.928285 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 09:03:29 crc kubenswrapper[4799]: I1124 09:03:29.934029 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.109166 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3b0342-d46b-42ef-90f8-c46be0e942b8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8e3b0342-d46b-42ef-90f8-c46be0e942b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.109244 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw9hv\" (UniqueName: \"kubernetes.io/projected/8e3b0342-d46b-42ef-90f8-c46be0e942b8-kube-api-access-jw9hv\") pod \"nova-cell1-conductor-0\" (UID: \"8e3b0342-d46b-42ef-90f8-c46be0e942b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.109407 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3b0342-d46b-42ef-90f8-c46be0e942b8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8e3b0342-d46b-42ef-90f8-c46be0e942b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.212290 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3b0342-d46b-42ef-90f8-c46be0e942b8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8e3b0342-d46b-42ef-90f8-c46be0e942b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.212840 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3b0342-d46b-42ef-90f8-c46be0e942b8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8e3b0342-d46b-42ef-90f8-c46be0e942b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.212882 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw9hv\" (UniqueName: \"kubernetes.io/projected/8e3b0342-d46b-42ef-90f8-c46be0e942b8-kube-api-access-jw9hv\") pod \"nova-cell1-conductor-0\" (UID: \"8e3b0342-d46b-42ef-90f8-c46be0e942b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.221987 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3b0342-d46b-42ef-90f8-c46be0e942b8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8e3b0342-d46b-42ef-90f8-c46be0e942b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.223192 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3b0342-d46b-42ef-90f8-c46be0e942b8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8e3b0342-d46b-42ef-90f8-c46be0e942b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.230303 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw9hv\" (UniqueName: \"kubernetes.io/projected/8e3b0342-d46b-42ef-90f8-c46be0e942b8-kube-api-access-jw9hv\") pod \"nova-cell1-conductor-0\" (UID: \"8e3b0342-d46b-42ef-90f8-c46be0e942b8\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.255345 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.499128 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.622678 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-config-data\") pod \"2516f20b-d128-4383-9982-699005340d35\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.623367 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-combined-ca-bundle\") pod \"2516f20b-d128-4383-9982-699005340d35\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.623454 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssvlh\" (UniqueName: \"kubernetes.io/projected/2516f20b-d128-4383-9982-699005340d35-kube-api-access-ssvlh\") pod \"2516f20b-d128-4383-9982-699005340d35\" (UID: \"2516f20b-d128-4383-9982-699005340d35\") " Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.629109 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2516f20b-d128-4383-9982-699005340d35-kube-api-access-ssvlh" (OuterVolumeSpecName: "kube-api-access-ssvlh") pod "2516f20b-d128-4383-9982-699005340d35" (UID: "2516f20b-d128-4383-9982-699005340d35"). InnerVolumeSpecName "kube-api-access-ssvlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.657011 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2516f20b-d128-4383-9982-699005340d35" (UID: "2516f20b-d128-4383-9982-699005340d35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.662694 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-config-data" (OuterVolumeSpecName: "config-data") pod "2516f20b-d128-4383-9982-699005340d35" (UID: "2516f20b-d128-4383-9982-699005340d35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.726395 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.726439 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2516f20b-d128-4383-9982-699005340d35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.726457 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssvlh\" (UniqueName: \"kubernetes.io/projected/2516f20b-d128-4383-9982-699005340d35-kube-api-access-ssvlh\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.819211 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:03:30 crc kubenswrapper[4799]: W1124 09:03:30.831603 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e3b0342_d46b_42ef_90f8_c46be0e942b8.slice/crio-42b9c38672d0c9d18bf9bc2aef5e5f5ec67c0b3da1005121ef51852d854aeb16 WatchSource:0}: Error finding container 42b9c38672d0c9d18bf9bc2aef5e5f5ec67c0b3da1005121ef51852d854aeb16: Status 404 returned error can't find the container with id 42b9c38672d0c9d18bf9bc2aef5e5f5ec67c0b3da1005121ef51852d854aeb16 Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.876328 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" event={"ID":"4ad610ec-2cdd-4eea-8d88-18b633d28225","Type":"ContainerStarted","Data":"5ba83a744d8bcc434a455476c2d4a25d27056821d8a7bb28bbfa4e21ddaeb4c3"} Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.877551 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8e3b0342-d46b-42ef-90f8-c46be0e942b8","Type":"ContainerStarted","Data":"42b9c38672d0c9d18bf9bc2aef5e5f5ec67c0b3da1005121ef51852d854aeb16"} Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.884556 4799 generic.go:334] "Generic (PLEG): container finished" podID="2516f20b-d128-4383-9982-699005340d35" containerID="75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286" exitCode=0 Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.884658 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.884676 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2516f20b-d128-4383-9982-699005340d35","Type":"ContainerDied","Data":"75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286"} Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.885055 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2516f20b-d128-4383-9982-699005340d35","Type":"ContainerDied","Data":"839e1daf35d8c2935d8a5f1329d56edab47fe9130467f0a56867b910bc467355"} Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.885098 4799 scope.go:117] "RemoveContainer" containerID="75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.899781 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" podStartSLOduration=2.456567006 podStartE2EDuration="2.899764764s" podCreationTimestamp="2025-11-24 09:03:28 +0000 UTC" firstStartedPulling="2025-11-24 09:03:29.461749833 +0000 UTC m=+8155.117732307" lastFinishedPulling="2025-11-24 09:03:29.904947591 +0000 UTC m=+8155.560930065" observedRunningTime="2025-11-24 09:03:30.894734712 +0000 UTC m=+8156.550717196" watchObservedRunningTime="2025-11-24 09:03:30.899764764 +0000 UTC m=+8156.555747238" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.932368 4799 scope.go:117] "RemoveContainer" containerID="75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.935551 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:03:30 crc kubenswrapper[4799]: E1124 09:03:30.935905 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286\": container with ID starting with 75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286 not found: ID does not exist" containerID="75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.935950 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286"} err="failed to get container status \"75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286\": rpc error: code = NotFound desc = could not find container \"75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286\": container with ID starting with 75151689045c39d80c8778a9be64924de64dcf17645f86b23df2b6d72aaca286 not found: ID does not exist" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.952048 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.969132 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:03:30 crc kubenswrapper[4799]: E1124 09:03:30.969659 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2516f20b-d128-4383-9982-699005340d35" containerName="nova-cell0-conductor-conductor" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.969678 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2516f20b-d128-4383-9982-699005340d35" containerName="nova-cell0-conductor-conductor" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.970022 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2516f20b-d128-4383-9982-699005340d35" containerName="nova-cell0-conductor-conductor" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.971121 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.973358 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 09:03:30 crc kubenswrapper[4799]: I1124 09:03:30.981731 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.135740 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m2tf\" (UniqueName: \"kubernetes.io/projected/4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6-kube-api-access-2m2tf\") pod \"nova-cell0-conductor-0\" (UID: \"4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.136103 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.136363 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.238281 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.238359 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m2tf\" (UniqueName: \"kubernetes.io/projected/4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6-kube-api-access-2m2tf\") pod \"nova-cell0-conductor-0\" (UID: \"4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.238450 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.243921 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.244677 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.257808 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m2tf\" (UniqueName: \"kubernetes.io/projected/4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6-kube-api-access-2m2tf\") pod \"nova-cell0-conductor-0\" (UID: \"4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.312994 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.521129 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": read tcp 10.217.0.2:42644->10.217.1.81:8775: read: connection reset by peer" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.521157 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": read tcp 10.217.0.2:42640->10.217.1.81:8775: read: connection reset by peer" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.647190 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2516f20b-d128-4383-9982-699005340d35" path="/var/lib/kubelet/pods/2516f20b-d128-4383-9982-699005340d35/volumes" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.647788 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ceaea61-e8ed-455e-8020-b96783d44e72" path="/var/lib/kubelet/pods/2ceaea61-e8ed-455e-8020-b96783d44e72/volumes" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.781192 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.919891 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19134f4a-118d-4e1c-b53d-742d563576c0","Type":"ContainerDied","Data":"4eca8afcb5ad58ce56463cb107cd2c212cce92b084a87e47d7e2c55e7954ea6c"} Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.919830 4799 generic.go:334] "Generic (PLEG): container finished" podID="19134f4a-118d-4e1c-b53d-742d563576c0" containerID="4eca8afcb5ad58ce56463cb107cd2c212cce92b084a87e47d7e2c55e7954ea6c" exitCode=0 Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.924825 4799 generic.go:334] "Generic (PLEG): container finished" podID="3966b7b5-c531-4127-82d0-8d759464708a" containerID="53b35860f548f35f4338efbdd33b9ee83820c2b24af2065c9e20cbe4f405183b" exitCode=0 Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.924891 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3966b7b5-c531-4127-82d0-8d759464708a","Type":"ContainerDied","Data":"53b35860f548f35f4338efbdd33b9ee83820c2b24af2065c9e20cbe4f405183b"} Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.924910 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3966b7b5-c531-4127-82d0-8d759464708a","Type":"ContainerDied","Data":"59c3bcdd109c1136c39115f7eb46b0a21614e3f1e2e2946d8e92b0a48d2b8a65"} Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.924920 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59c3bcdd109c1136c39115f7eb46b0a21614e3f1e2e2946d8e92b0a48d2b8a65" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.944117 4799 generic.go:334] "Generic (PLEG): container finished" podID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerID="bae71ae1cf9e4b2ec05692bd4e70b2a8132ec02f03bac45d32f8a3fe145198a9" exitCode=0 Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.944181 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27fb95fd-a5be-4c99-9ce3-f8617695a530","Type":"ContainerDied","Data":"bae71ae1cf9e4b2ec05692bd4e70b2a8132ec02f03bac45d32f8a3fe145198a9"} Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.948518 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8e3b0342-d46b-42ef-90f8-c46be0e942b8","Type":"ContainerStarted","Data":"de2f0c0fe8150404a8bb084d23bf39069f86662cc0c117341c6300837d81e9fd"} Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.949588 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.967779 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6","Type":"ContainerStarted","Data":"021debc39d0036cc453bc47a2f3c63fcba024f50cd67d5340948fb8fefc5300f"} Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.972308 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.972286317 podStartE2EDuration="2.972286317s" podCreationTimestamp="2025-11-24 09:03:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:03:31.966900775 +0000 UTC m=+8157.622883259" watchObservedRunningTime="2025-11-24 09:03:31.972286317 +0000 UTC m=+8157.628268791" Nov 24 09:03:31 crc kubenswrapper[4799]: I1124 09:03:31.990029 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.032886 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.173338 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-config-data\") pod \"3966b7b5-c531-4127-82d0-8d759464708a\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.173415 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lms8v\" (UniqueName: \"kubernetes.io/projected/3966b7b5-c531-4127-82d0-8d759464708a-kube-api-access-lms8v\") pod \"3966b7b5-c531-4127-82d0-8d759464708a\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.173652 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvjcp\" (UniqueName: \"kubernetes.io/projected/27fb95fd-a5be-4c99-9ce3-f8617695a530-kube-api-access-nvjcp\") pod \"27fb95fd-a5be-4c99-9ce3-f8617695a530\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.173697 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-combined-ca-bundle\") pod \"27fb95fd-a5be-4c99-9ce3-f8617695a530\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.173836 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-combined-ca-bundle\") pod \"3966b7b5-c531-4127-82d0-8d759464708a\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.173913 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3966b7b5-c531-4127-82d0-8d759464708a-logs\") pod \"3966b7b5-c531-4127-82d0-8d759464708a\" (UID: \"3966b7b5-c531-4127-82d0-8d759464708a\") " Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.174004 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27fb95fd-a5be-4c99-9ce3-f8617695a530-logs\") pod \"27fb95fd-a5be-4c99-9ce3-f8617695a530\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.174107 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-config-data\") pod \"27fb95fd-a5be-4c99-9ce3-f8617695a530\" (UID: \"27fb95fd-a5be-4c99-9ce3-f8617695a530\") " Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.175556 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3966b7b5-c531-4127-82d0-8d759464708a-logs" (OuterVolumeSpecName: "logs") pod "3966b7b5-c531-4127-82d0-8d759464708a" (UID: "3966b7b5-c531-4127-82d0-8d759464708a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.176006 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27fb95fd-a5be-4c99-9ce3-f8617695a530-logs" (OuterVolumeSpecName: "logs") pod "27fb95fd-a5be-4c99-9ce3-f8617695a530" (UID: "27fb95fd-a5be-4c99-9ce3-f8617695a530"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.177626 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3966b7b5-c531-4127-82d0-8d759464708a-kube-api-access-lms8v" (OuterVolumeSpecName: "kube-api-access-lms8v") pod "3966b7b5-c531-4127-82d0-8d759464708a" (UID: "3966b7b5-c531-4127-82d0-8d759464708a"). InnerVolumeSpecName "kube-api-access-lms8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.180583 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27fb95fd-a5be-4c99-9ce3-f8617695a530-kube-api-access-nvjcp" (OuterVolumeSpecName: "kube-api-access-nvjcp") pod "27fb95fd-a5be-4c99-9ce3-f8617695a530" (UID: "27fb95fd-a5be-4c99-9ce3-f8617695a530"). InnerVolumeSpecName "kube-api-access-nvjcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.206317 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-config-data" (OuterVolumeSpecName: "config-data") pod "3966b7b5-c531-4127-82d0-8d759464708a" (UID: "3966b7b5-c531-4127-82d0-8d759464708a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.207960 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-config-data" (OuterVolumeSpecName: "config-data") pod "27fb95fd-a5be-4c99-9ce3-f8617695a530" (UID: "27fb95fd-a5be-4c99-9ce3-f8617695a530"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.221077 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3966b7b5-c531-4127-82d0-8d759464708a" (UID: "3966b7b5-c531-4127-82d0-8d759464708a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.222556 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27fb95fd-a5be-4c99-9ce3-f8617695a530" (UID: "27fb95fd-a5be-4c99-9ce3-f8617695a530"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.264170 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.277301 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvjcp\" (UniqueName: \"kubernetes.io/projected/27fb95fd-a5be-4c99-9ce3-f8617695a530-kube-api-access-nvjcp\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.277338 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.277353 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.277365 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3966b7b5-c531-4127-82d0-8d759464708a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.277379 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27fb95fd-a5be-4c99-9ce3-f8617695a530-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.277390 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27fb95fd-a5be-4c99-9ce3-f8617695a530-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.277404 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3966b7b5-c531-4127-82d0-8d759464708a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.277416 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lms8v\" (UniqueName: \"kubernetes.io/projected/3966b7b5-c531-4127-82d0-8d759464708a-kube-api-access-lms8v\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.378308 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-config-data\") pod \"19134f4a-118d-4e1c-b53d-742d563576c0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.378392 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-combined-ca-bundle\") pod \"19134f4a-118d-4e1c-b53d-742d563576c0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.378682 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhgzs\" (UniqueName: \"kubernetes.io/projected/19134f4a-118d-4e1c-b53d-742d563576c0-kube-api-access-dhgzs\") pod \"19134f4a-118d-4e1c-b53d-742d563576c0\" (UID: \"19134f4a-118d-4e1c-b53d-742d563576c0\") " Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.385390 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19134f4a-118d-4e1c-b53d-742d563576c0-kube-api-access-dhgzs" (OuterVolumeSpecName: "kube-api-access-dhgzs") pod "19134f4a-118d-4e1c-b53d-742d563576c0" (UID: "19134f4a-118d-4e1c-b53d-742d563576c0"). InnerVolumeSpecName "kube-api-access-dhgzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.423163 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19134f4a-118d-4e1c-b53d-742d563576c0" (UID: "19134f4a-118d-4e1c-b53d-742d563576c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.431006 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-config-data" (OuterVolumeSpecName: "config-data") pod "19134f4a-118d-4e1c-b53d-742d563576c0" (UID: "19134f4a-118d-4e1c-b53d-742d563576c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.481252 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhgzs\" (UniqueName: \"kubernetes.io/projected/19134f4a-118d-4e1c-b53d-742d563576c0-kube-api-access-dhgzs\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.481298 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.481312 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19134f4a-118d-4e1c-b53d-742d563576c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.977802 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6","Type":"ContainerStarted","Data":"1f3f21c6b5a83d1b1043c7561eff735a3170030f93bd48a30a7fa86eab33e8c8"} Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.978273 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.981190 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19134f4a-118d-4e1c-b53d-742d563576c0","Type":"ContainerDied","Data":"319e8891665d9dd4f986476ebbf101d669103d15f4b7224aa2d2cf6712aff7db"} Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.981244 4799 scope.go:117] "RemoveContainer" containerID="4eca8afcb5ad58ce56463cb107cd2c212cce92b084a87e47d7e2c55e7954ea6c" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.981402 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.986583 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.986695 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:03:32 crc kubenswrapper[4799]: I1124 09:03:32.986599 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27fb95fd-a5be-4c99-9ce3-f8617695a530","Type":"ContainerDied","Data":"f19e33828d5e640ad17a5d1eaa9c2924d4a5d2d0adad9ffae4eeebd7b0448b17"} Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.012227 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.012210335 podStartE2EDuration="3.012210335s" podCreationTimestamp="2025-11-24 09:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:03:33.008243485 +0000 UTC m=+8158.664225959" watchObservedRunningTime="2025-11-24 09:03:33.012210335 +0000 UTC m=+8158.668192809" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.017215 4799 scope.go:117] "RemoveContainer" containerID="bae71ae1cf9e4b2ec05692bd4e70b2a8132ec02f03bac45d32f8a3fe145198a9" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.043653 4799 scope.go:117] "RemoveContainer" containerID="de6e817697e87c1221c2c1aa6845284ecd7b05c5c743fa80dc4ce0ed79f8fbc0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.047925 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.063985 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.078455 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: E1124 09:03:33.078979 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3966b7b5-c531-4127-82d0-8d759464708a" containerName="nova-api-api" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.078998 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3966b7b5-c531-4127-82d0-8d759464708a" containerName="nova-api-api" Nov 24 09:03:33 crc kubenswrapper[4799]: E1124 09:03:33.079016 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-log" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.079023 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-log" Nov 24 09:03:33 crc kubenswrapper[4799]: E1124 09:03:33.079053 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3966b7b5-c531-4127-82d0-8d759464708a" containerName="nova-api-log" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.079059 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3966b7b5-c531-4127-82d0-8d759464708a" containerName="nova-api-log" Nov 24 09:03:33 crc kubenswrapper[4799]: E1124 09:03:33.079072 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19134f4a-118d-4e1c-b53d-742d563576c0" containerName="nova-scheduler-scheduler" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.079078 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="19134f4a-118d-4e1c-b53d-742d563576c0" containerName="nova-scheduler-scheduler" Nov 24 09:03:33 crc kubenswrapper[4799]: E1124 09:03:33.079097 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-metadata" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.079104 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-metadata" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.079311 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-metadata" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.079326 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="19134f4a-118d-4e1c-b53d-742d563576c0" containerName="nova-scheduler-scheduler" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.079336 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3966b7b5-c531-4127-82d0-8d759464708a" containerName="nova-api-log" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.079352 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3966b7b5-c531-4127-82d0-8d759464708a" containerName="nova-api-api" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.079364 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" containerName="nova-metadata-log" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.080135 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.097115 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.109308 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.118994 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.136721 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.145434 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.158087 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.168920 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.171546 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.173972 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.176892 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.185363 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.189358 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.191635 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.194141 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.202956 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17aeb60b-35e6-4096-bdce-26ac7c14797c-config-data\") pod \"nova-scheduler-0\" (UID: \"17aeb60b-35e6-4096-bdce-26ac7c14797c\") " pod="openstack/nova-scheduler-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.203012 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17aeb60b-35e6-4096-bdce-26ac7c14797c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"17aeb60b-35e6-4096-bdce-26ac7c14797c\") " pod="openstack/nova-scheduler-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.203181 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvzvw\" (UniqueName: \"kubernetes.io/projected/17aeb60b-35e6-4096-bdce-26ac7c14797c-kube-api-access-tvzvw\") pod \"nova-scheduler-0\" (UID: \"17aeb60b-35e6-4096-bdce-26ac7c14797c\") " pod="openstack/nova-scheduler-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.304788 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btndp\" (UniqueName: \"kubernetes.io/projected/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-kube-api-access-btndp\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.304876 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvzvw\" (UniqueName: \"kubernetes.io/projected/17aeb60b-35e6-4096-bdce-26ac7c14797c-kube-api-access-tvzvw\") pod \"nova-scheduler-0\" (UID: \"17aeb60b-35e6-4096-bdce-26ac7c14797c\") " pod="openstack/nova-scheduler-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.304900 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28psf\" (UniqueName: \"kubernetes.io/projected/2238727c-e84a-4641-b168-6168ec24a46a-kube-api-access-28psf\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.304951 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-config-data\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.304984 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-logs\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.305157 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2238727c-e84a-4641-b168-6168ec24a46a-config-data\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.305205 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17aeb60b-35e6-4096-bdce-26ac7c14797c-config-data\") pod \"nova-scheduler-0\" (UID: \"17aeb60b-35e6-4096-bdce-26ac7c14797c\") " pod="openstack/nova-scheduler-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.305224 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17aeb60b-35e6-4096-bdce-26ac7c14797c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"17aeb60b-35e6-4096-bdce-26ac7c14797c\") " pod="openstack/nova-scheduler-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.305347 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2238727c-e84a-4641-b168-6168ec24a46a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.305534 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2238727c-e84a-4641-b168-6168ec24a46a-logs\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.305603 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.310149 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17aeb60b-35e6-4096-bdce-26ac7c14797c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"17aeb60b-35e6-4096-bdce-26ac7c14797c\") " pod="openstack/nova-scheduler-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.310233 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17aeb60b-35e6-4096-bdce-26ac7c14797c-config-data\") pod \"nova-scheduler-0\" (UID: \"17aeb60b-35e6-4096-bdce-26ac7c14797c\") " pod="openstack/nova-scheduler-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.322374 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvzvw\" (UniqueName: \"kubernetes.io/projected/17aeb60b-35e6-4096-bdce-26ac7c14797c-kube-api-access-tvzvw\") pod \"nova-scheduler-0\" (UID: \"17aeb60b-35e6-4096-bdce-26ac7c14797c\") " pod="openstack/nova-scheduler-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.407131 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btndp\" (UniqueName: \"kubernetes.io/projected/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-kube-api-access-btndp\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.407198 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28psf\" (UniqueName: \"kubernetes.io/projected/2238727c-e84a-4641-b168-6168ec24a46a-kube-api-access-28psf\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.407257 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-config-data\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.407297 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-logs\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.407338 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2238727c-e84a-4641-b168-6168ec24a46a-config-data\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.407439 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2238727c-e84a-4641-b168-6168ec24a46a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.407496 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2238727c-e84a-4641-b168-6168ec24a46a-logs\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.407526 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.408105 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-logs\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.408345 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2238727c-e84a-4641-b168-6168ec24a46a-logs\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.415461 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2238727c-e84a-4641-b168-6168ec24a46a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.415501 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-config-data\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.415866 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2238727c-e84a-4641-b168-6168ec24a46a-config-data\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.420017 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.421466 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.436706 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28psf\" (UniqueName: \"kubernetes.io/projected/2238727c-e84a-4641-b168-6168ec24a46a-kube-api-access-28psf\") pod \"nova-metadata-0\" (UID: \"2238727c-e84a-4641-b168-6168ec24a46a\") " pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.437351 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btndp\" (UniqueName: \"kubernetes.io/projected/0863a6f5-d8a7-4c50-a33b-3398e262a9e7-kube-api-access-btndp\") pod \"nova-api-0\" (UID: \"0863a6f5-d8a7-4c50-a33b-3398e262a9e7\") " pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.490455 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.512434 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.652370 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19134f4a-118d-4e1c-b53d-742d563576c0" path="/var/lib/kubelet/pods/19134f4a-118d-4e1c-b53d-742d563576c0/volumes" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.656770 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27fb95fd-a5be-4c99-9ce3-f8617695a530" path="/var/lib/kubelet/pods/27fb95fd-a5be-4c99-9ce3-f8617695a530/volumes" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.658039 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3966b7b5-c531-4127-82d0-8d759464708a" path="/var/lib/kubelet/pods/3966b7b5-c531-4127-82d0-8d759464708a/volumes" Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.882549 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:03:33 crc kubenswrapper[4799]: I1124 09:03:33.998137 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"17aeb60b-35e6-4096-bdce-26ac7c14797c","Type":"ContainerStarted","Data":"f421db9a5d441ccdddefbc5b72df68665dc3a7add0c0d483dc40ca96adf9adc9"} Nov 24 09:03:34 crc kubenswrapper[4799]: I1124 09:03:34.066108 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:03:34 crc kubenswrapper[4799]: W1124 09:03:34.071257 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0863a6f5_d8a7_4c50_a33b_3398e262a9e7.slice/crio-699e18a4ebffbfc3242c94a5885109fc4b8f3d89d5e89124233b3dd713440876 WatchSource:0}: Error finding container 699e18a4ebffbfc3242c94a5885109fc4b8f3d89d5e89124233b3dd713440876: Status 404 returned error can't find the container with id 699e18a4ebffbfc3242c94a5885109fc4b8f3d89d5e89124233b3dd713440876 Nov 24 09:03:34 crc kubenswrapper[4799]: W1124 09:03:34.075472 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2238727c_e84a_4641_b168_6168ec24a46a.slice/crio-e38252b19eb8bb7624afa33f83521f2a3aeceead49c9644b5acb3f0f6bda11d3 WatchSource:0}: Error finding container e38252b19eb8bb7624afa33f83521f2a3aeceead49c9644b5acb3f0f6bda11d3: Status 404 returned error can't find the container with id e38252b19eb8bb7624afa33f83521f2a3aeceead49c9644b5acb3f0f6bda11d3 Nov 24 09:03:34 crc kubenswrapper[4799]: I1124 09:03:34.076175 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:03:35 crc kubenswrapper[4799]: I1124 09:03:35.021461 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"17aeb60b-35e6-4096-bdce-26ac7c14797c","Type":"ContainerStarted","Data":"baba0d36316d57b7d4329fd3dd5af43384f6d2e6fa8477342bae44506e5c76b9"} Nov 24 09:03:35 crc kubenswrapper[4799]: I1124 09:03:35.026351 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0863a6f5-d8a7-4c50-a33b-3398e262a9e7","Type":"ContainerStarted","Data":"207df81d846fd4f7eaf77dcd05f04b3ac3046e3b78d0609f021878997d0171d5"} Nov 24 09:03:35 crc kubenswrapper[4799]: I1124 09:03:35.026415 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0863a6f5-d8a7-4c50-a33b-3398e262a9e7","Type":"ContainerStarted","Data":"17f5f25c09398118219b1b8758a95d83dfca47cee6a405f8a9aaadec116adcb6"} Nov 24 09:03:35 crc kubenswrapper[4799]: I1124 09:03:35.026429 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0863a6f5-d8a7-4c50-a33b-3398e262a9e7","Type":"ContainerStarted","Data":"699e18a4ebffbfc3242c94a5885109fc4b8f3d89d5e89124233b3dd713440876"} Nov 24 09:03:35 crc kubenswrapper[4799]: I1124 09:03:35.028553 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2238727c-e84a-4641-b168-6168ec24a46a","Type":"ContainerStarted","Data":"103aeeff1c1dbe1399acbe99b700b2c2c2cd8030338b22c818ce240481ce9e9d"} Nov 24 09:03:35 crc kubenswrapper[4799]: I1124 09:03:35.028580 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2238727c-e84a-4641-b168-6168ec24a46a","Type":"ContainerStarted","Data":"b15674879b1c8a149682b52bd6fc9bb318fd2fc94df3b102633ff0a1f0fa2237"} Nov 24 09:03:35 crc kubenswrapper[4799]: I1124 09:03:35.028590 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2238727c-e84a-4641-b168-6168ec24a46a","Type":"ContainerStarted","Data":"e38252b19eb8bb7624afa33f83521f2a3aeceead49c9644b5acb3f0f6bda11d3"} Nov 24 09:03:35 crc kubenswrapper[4799]: I1124 09:03:35.037814 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.037784979 podStartE2EDuration="2.037784979s" podCreationTimestamp="2025-11-24 09:03:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:03:35.035039866 +0000 UTC m=+8160.691022340" watchObservedRunningTime="2025-11-24 09:03:35.037784979 +0000 UTC m=+8160.693767443" Nov 24 09:03:35 crc kubenswrapper[4799]: I1124 09:03:35.054425 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.054394631 podStartE2EDuration="2.054394631s" podCreationTimestamp="2025-11-24 09:03:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:03:35.05072502 +0000 UTC m=+8160.706707494" watchObservedRunningTime="2025-11-24 09:03:35.054394631 +0000 UTC m=+8160.710377105" Nov 24 09:03:35 crc kubenswrapper[4799]: I1124 09:03:35.073275 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.073256711 podStartE2EDuration="2.073256711s" podCreationTimestamp="2025-11-24 09:03:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:03:35.071198089 +0000 UTC m=+8160.727180563" watchObservedRunningTime="2025-11-24 09:03:35.073256711 +0000 UTC m=+8160.729239185" Nov 24 09:03:38 crc kubenswrapper[4799]: I1124 09:03:38.420505 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 09:03:38 crc kubenswrapper[4799]: I1124 09:03:38.513326 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 09:03:38 crc kubenswrapper[4799]: I1124 09:03:38.513374 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 09:03:40 crc kubenswrapper[4799]: I1124 09:03:40.283736 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 09:03:41 crc kubenswrapper[4799]: I1124 09:03:41.345806 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 09:03:43 crc kubenswrapper[4799]: I1124 09:03:43.420509 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 09:03:43 crc kubenswrapper[4799]: I1124 09:03:43.462556 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 09:03:43 crc kubenswrapper[4799]: I1124 09:03:43.491597 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 09:03:43 crc kubenswrapper[4799]: I1124 09:03:43.491636 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 09:03:43 crc kubenswrapper[4799]: I1124 09:03:43.514607 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 09:03:43 crc kubenswrapper[4799]: I1124 09:03:43.514646 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 09:03:44 crc kubenswrapper[4799]: I1124 09:03:44.146501 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 09:03:44 crc kubenswrapper[4799]: I1124 09:03:44.656121 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2238727c-e84a-4641-b168-6168ec24a46a" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.176:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 09:03:44 crc kubenswrapper[4799]: I1124 09:03:44.656223 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0863a6f5-d8a7-4c50-a33b-3398e262a9e7" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.175:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 09:03:44 crc kubenswrapper[4799]: I1124 09:03:44.656358 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0863a6f5-d8a7-4c50-a33b-3398e262a9e7" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.175:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 09:03:44 crc kubenswrapper[4799]: I1124 09:03:44.656411 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2238727c-e84a-4641-b168-6168ec24a46a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.176:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 09:03:50 crc kubenswrapper[4799]: I1124 09:03:50.400387 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:03:50 crc kubenswrapper[4799]: I1124 09:03:50.400900 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:03:53 crc kubenswrapper[4799]: I1124 09:03:53.495208 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 09:03:53 crc kubenswrapper[4799]: I1124 09:03:53.496747 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 09:03:53 crc kubenswrapper[4799]: I1124 09:03:53.497173 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 09:03:53 crc kubenswrapper[4799]: I1124 09:03:53.497196 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 09:03:53 crc kubenswrapper[4799]: I1124 09:03:53.500085 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 09:03:53 crc kubenswrapper[4799]: I1124 09:03:53.500269 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 09:03:53 crc kubenswrapper[4799]: I1124 09:03:53.519160 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 09:03:53 crc kubenswrapper[4799]: I1124 09:03:53.520274 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 09:03:53 crc kubenswrapper[4799]: I1124 09:03:53.523202 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 09:03:54 crc kubenswrapper[4799]: I1124 09:03:54.231498 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 09:04:12 crc kubenswrapper[4799]: I1124 09:04:12.126243 4799 scope.go:117] "RemoveContainer" containerID="46251192c28f8377b93fd73ac7b5fda79e3f27f7ede96780c4906f47d9f1d9e3" Nov 24 09:04:12 crc kubenswrapper[4799]: I1124 09:04:12.152565 4799 scope.go:117] "RemoveContainer" containerID="53b35860f548f35f4338efbdd33b9ee83820c2b24af2065c9e20cbe4f405183b" Nov 24 09:04:20 crc kubenswrapper[4799]: I1124 09:04:20.400308 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:04:20 crc kubenswrapper[4799]: I1124 09:04:20.400948 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:04:49 crc kubenswrapper[4799]: I1124 09:04:49.930311 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cfbqx"] Nov 24 09:04:49 crc kubenswrapper[4799]: I1124 09:04:49.935196 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:04:49 crc kubenswrapper[4799]: I1124 09:04:49.952316 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cfbqx"] Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.114590 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-catalog-content\") pod \"community-operators-cfbqx\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.114711 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hthrn\" (UniqueName: \"kubernetes.io/projected/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-kube-api-access-hthrn\") pod \"community-operators-cfbqx\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.114758 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-utilities\") pod \"community-operators-cfbqx\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.216595 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hthrn\" (UniqueName: \"kubernetes.io/projected/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-kube-api-access-hthrn\") pod \"community-operators-cfbqx\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.216696 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-utilities\") pod \"community-operators-cfbqx\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.217152 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-utilities\") pod \"community-operators-cfbqx\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.217326 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-catalog-content\") pod \"community-operators-cfbqx\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.217615 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-catalog-content\") pod \"community-operators-cfbqx\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.235545 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hthrn\" (UniqueName: \"kubernetes.io/projected/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-kube-api-access-hthrn\") pod \"community-operators-cfbqx\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.256559 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.402383 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.402733 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.402785 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.403863 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c892fde2c41f62dc151886b156c9c281cd73f67e86c585ace87fe5a19ac868d"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.403931 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://6c892fde2c41f62dc151886b156c9c281cd73f67e86c585ace87fe5a19ac868d" gracePeriod=600 Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.748193 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cfbqx"] Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.837386 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfbqx" event={"ID":"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56","Type":"ContainerStarted","Data":"4a5468f56d8e84cb8a76f8eb22e861f5f3042ffa3b618668503b235d22b8b4cb"} Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.884741 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="6c892fde2c41f62dc151886b156c9c281cd73f67e86c585ace87fe5a19ac868d" exitCode=0 Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.884831 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"6c892fde2c41f62dc151886b156c9c281cd73f67e86c585ace87fe5a19ac868d"} Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.884892 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e"} Nov 24 09:04:50 crc kubenswrapper[4799]: I1124 09:04:50.885122 4799 scope.go:117] "RemoveContainer" containerID="75e8508499632084a6d7d0eee1f90f604aa79c49b8549b2e1ebd6db90ad4170b" Nov 24 09:04:51 crc kubenswrapper[4799]: I1124 09:04:51.899526 4799 generic.go:334] "Generic (PLEG): container finished" podID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" containerID="f2c0fa6b6bc00026909e8931065b352851ca3eaf87728d5c09fa2807538ce19c" exitCode=0 Nov 24 09:04:51 crc kubenswrapper[4799]: I1124 09:04:51.899594 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfbqx" event={"ID":"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56","Type":"ContainerDied","Data":"f2c0fa6b6bc00026909e8931065b352851ca3eaf87728d5c09fa2807538ce19c"} Nov 24 09:04:51 crc kubenswrapper[4799]: I1124 09:04:51.902143 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:04:53 crc kubenswrapper[4799]: I1124 09:04:53.922649 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfbqx" event={"ID":"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56","Type":"ContainerStarted","Data":"381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7"} Nov 24 09:04:54 crc kubenswrapper[4799]: I1124 09:04:54.934465 4799 generic.go:334] "Generic (PLEG): container finished" podID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" containerID="381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7" exitCode=0 Nov 24 09:04:54 crc kubenswrapper[4799]: I1124 09:04:54.934828 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfbqx" event={"ID":"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56","Type":"ContainerDied","Data":"381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7"} Nov 24 09:04:55 crc kubenswrapper[4799]: I1124 09:04:55.948998 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfbqx" event={"ID":"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56","Type":"ContainerStarted","Data":"d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1"} Nov 24 09:04:55 crc kubenswrapper[4799]: I1124 09:04:55.974358 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cfbqx" podStartSLOduration=3.331397635 podStartE2EDuration="6.974333019s" podCreationTimestamp="2025-11-24 09:04:49 +0000 UTC" firstStartedPulling="2025-11-24 09:04:51.901813615 +0000 UTC m=+8237.557796089" lastFinishedPulling="2025-11-24 09:04:55.544748999 +0000 UTC m=+8241.200731473" observedRunningTime="2025-11-24 09:04:55.972568399 +0000 UTC m=+8241.628550873" watchObservedRunningTime="2025-11-24 09:04:55.974333019 +0000 UTC m=+8241.630315493" Nov 24 09:05:00 crc kubenswrapper[4799]: I1124 09:05:00.257419 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:05:00 crc kubenswrapper[4799]: I1124 09:05:00.258056 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:05:00 crc kubenswrapper[4799]: I1124 09:05:00.306777 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:05:01 crc kubenswrapper[4799]: I1124 09:05:01.062400 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:05:02 crc kubenswrapper[4799]: I1124 09:05:02.518287 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cfbqx"] Nov 24 09:05:03 crc kubenswrapper[4799]: I1124 09:05:03.028918 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cfbqx" podUID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" containerName="registry-server" containerID="cri-o://d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1" gracePeriod=2 Nov 24 09:05:03 crc kubenswrapper[4799]: I1124 09:05:03.521721 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:05:03 crc kubenswrapper[4799]: I1124 09:05:03.648066 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-utilities\") pod \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " Nov 24 09:05:03 crc kubenswrapper[4799]: I1124 09:05:03.648376 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hthrn\" (UniqueName: \"kubernetes.io/projected/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-kube-api-access-hthrn\") pod \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " Nov 24 09:05:03 crc kubenswrapper[4799]: I1124 09:05:03.648448 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-catalog-content\") pod \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\" (UID: \"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56\") " Nov 24 09:05:03 crc kubenswrapper[4799]: I1124 09:05:03.652729 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-utilities" (OuterVolumeSpecName: "utilities") pod "ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" (UID: "ba5bcd21-d8f6-4c31-ab87-57ac5a345f56"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:05:03 crc kubenswrapper[4799]: I1124 09:05:03.659722 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-kube-api-access-hthrn" (OuterVolumeSpecName: "kube-api-access-hthrn") pod "ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" (UID: "ba5bcd21-d8f6-4c31-ab87-57ac5a345f56"). InnerVolumeSpecName "kube-api-access-hthrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:05:03 crc kubenswrapper[4799]: I1124 09:05:03.751223 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hthrn\" (UniqueName: \"kubernetes.io/projected/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-kube-api-access-hthrn\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:03 crc kubenswrapper[4799]: I1124 09:05:03.751571 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.059042 4799 generic.go:334] "Generic (PLEG): container finished" podID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" containerID="d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1" exitCode=0 Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.059144 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfbqx" event={"ID":"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56","Type":"ContainerDied","Data":"d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1"} Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.059205 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfbqx" event={"ID":"ba5bcd21-d8f6-4c31-ab87-57ac5a345f56","Type":"ContainerDied","Data":"4a5468f56d8e84cb8a76f8eb22e861f5f3042ffa3b618668503b235d22b8b4cb"} Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.059237 4799 scope.go:117] "RemoveContainer" containerID="d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.059315 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfbqx" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.080448 4799 scope.go:117] "RemoveContainer" containerID="381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.102152 4799 scope.go:117] "RemoveContainer" containerID="f2c0fa6b6bc00026909e8931065b352851ca3eaf87728d5c09fa2807538ce19c" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.147908 4799 scope.go:117] "RemoveContainer" containerID="d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1" Nov 24 09:05:04 crc kubenswrapper[4799]: E1124 09:05:04.148329 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1\": container with ID starting with d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1 not found: ID does not exist" containerID="d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.148361 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1"} err="failed to get container status \"d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1\": rpc error: code = NotFound desc = could not find container \"d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1\": container with ID starting with d44ecc4170d6fc3ae500194babfb428b0b387cfcb102d4fe6f8a12b4d28869f1 not found: ID does not exist" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.148381 4799 scope.go:117] "RemoveContainer" containerID="381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7" Nov 24 09:05:04 crc kubenswrapper[4799]: E1124 09:05:04.148583 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7\": container with ID starting with 381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7 not found: ID does not exist" containerID="381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.148604 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7"} err="failed to get container status \"381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7\": rpc error: code = NotFound desc = could not find container \"381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7\": container with ID starting with 381fe53ca842a2ba1846b6a54b6af51fc5232e4e9db0fb9cd2563bba34569ea7 not found: ID does not exist" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.148618 4799 scope.go:117] "RemoveContainer" containerID="f2c0fa6b6bc00026909e8931065b352851ca3eaf87728d5c09fa2807538ce19c" Nov 24 09:05:04 crc kubenswrapper[4799]: E1124 09:05:04.148812 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2c0fa6b6bc00026909e8931065b352851ca3eaf87728d5c09fa2807538ce19c\": container with ID starting with f2c0fa6b6bc00026909e8931065b352851ca3eaf87728d5c09fa2807538ce19c not found: ID does not exist" containerID="f2c0fa6b6bc00026909e8931065b352851ca3eaf87728d5c09fa2807538ce19c" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.148831 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2c0fa6b6bc00026909e8931065b352851ca3eaf87728d5c09fa2807538ce19c"} err="failed to get container status \"f2c0fa6b6bc00026909e8931065b352851ca3eaf87728d5c09fa2807538ce19c\": rpc error: code = NotFound desc = could not find container \"f2c0fa6b6bc00026909e8931065b352851ca3eaf87728d5c09fa2807538ce19c\": container with ID starting with f2c0fa6b6bc00026909e8931065b352851ca3eaf87728d5c09fa2807538ce19c not found: ID does not exist" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.466521 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" (UID: "ba5bcd21-d8f6-4c31-ab87-57ac5a345f56"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.470630 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.695331 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cfbqx"] Nov 24 09:05:04 crc kubenswrapper[4799]: I1124 09:05:04.704437 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cfbqx"] Nov 24 09:05:05 crc kubenswrapper[4799]: I1124 09:05:05.668554 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" path="/var/lib/kubelet/pods/ba5bcd21-d8f6-4c31-ab87-57ac5a345f56/volumes" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.003669 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tslvs"] Nov 24 09:06:44 crc kubenswrapper[4799]: E1124 09:06:44.004552 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" containerName="registry-server" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.004564 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" containerName="registry-server" Nov 24 09:06:44 crc kubenswrapper[4799]: E1124 09:06:44.004577 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" containerName="extract-utilities" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.004585 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" containerName="extract-utilities" Nov 24 09:06:44 crc kubenswrapper[4799]: E1124 09:06:44.004607 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" containerName="extract-content" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.004613 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" containerName="extract-content" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.004821 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba5bcd21-d8f6-4c31-ab87-57ac5a345f56" containerName="registry-server" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.006224 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.018521 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tslvs"] Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.150340 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-utilities\") pod \"certified-operators-tslvs\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.150455 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-catalog-content\") pod \"certified-operators-tslvs\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.150546 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvdcp\" (UniqueName: \"kubernetes.io/projected/99702864-6a6c-4f35-ae70-47a214c4c092-kube-api-access-xvdcp\") pod \"certified-operators-tslvs\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.253590 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvdcp\" (UniqueName: \"kubernetes.io/projected/99702864-6a6c-4f35-ae70-47a214c4c092-kube-api-access-xvdcp\") pod \"certified-operators-tslvs\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.254058 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-utilities\") pod \"certified-operators-tslvs\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.254151 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-catalog-content\") pod \"certified-operators-tslvs\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.254649 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-utilities\") pod \"certified-operators-tslvs\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.254707 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-catalog-content\") pod \"certified-operators-tslvs\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.276825 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvdcp\" (UniqueName: \"kubernetes.io/projected/99702864-6a6c-4f35-ae70-47a214c4c092-kube-api-access-xvdcp\") pod \"certified-operators-tslvs\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.335595 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:44 crc kubenswrapper[4799]: I1124 09:06:44.891613 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tslvs"] Nov 24 09:06:44 crc kubenswrapper[4799]: W1124 09:06:44.905993 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99702864_6a6c_4f35_ae70_47a214c4c092.slice/crio-9a28f6696d6bdeca606aa7c0b22b8142f8e44e32bd6484421071f99366e6af45 WatchSource:0}: Error finding container 9a28f6696d6bdeca606aa7c0b22b8142f8e44e32bd6484421071f99366e6af45: Status 404 returned error can't find the container with id 9a28f6696d6bdeca606aa7c0b22b8142f8e44e32bd6484421071f99366e6af45 Nov 24 09:06:45 crc kubenswrapper[4799]: I1124 09:06:45.752707 4799 generic.go:334] "Generic (PLEG): container finished" podID="99702864-6a6c-4f35-ae70-47a214c4c092" containerID="726c2410b3df89eba5bcc20f9c0fc472f2595833b53b6818c51dc366823bcf00" exitCode=0 Nov 24 09:06:45 crc kubenswrapper[4799]: I1124 09:06:45.754070 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tslvs" event={"ID":"99702864-6a6c-4f35-ae70-47a214c4c092","Type":"ContainerDied","Data":"726c2410b3df89eba5bcc20f9c0fc472f2595833b53b6818c51dc366823bcf00"} Nov 24 09:06:45 crc kubenswrapper[4799]: I1124 09:06:45.754195 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tslvs" event={"ID":"99702864-6a6c-4f35-ae70-47a214c4c092","Type":"ContainerStarted","Data":"9a28f6696d6bdeca606aa7c0b22b8142f8e44e32bd6484421071f99366e6af45"} Nov 24 09:06:47 crc kubenswrapper[4799]: I1124 09:06:47.786545 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tslvs" event={"ID":"99702864-6a6c-4f35-ae70-47a214c4c092","Type":"ContainerStarted","Data":"280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf"} Nov 24 09:06:48 crc kubenswrapper[4799]: I1124 09:06:48.800445 4799 generic.go:334] "Generic (PLEG): container finished" podID="99702864-6a6c-4f35-ae70-47a214c4c092" containerID="280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf" exitCode=0 Nov 24 09:06:48 crc kubenswrapper[4799]: I1124 09:06:48.800626 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tslvs" event={"ID":"99702864-6a6c-4f35-ae70-47a214c4c092","Type":"ContainerDied","Data":"280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf"} Nov 24 09:06:49 crc kubenswrapper[4799]: I1124 09:06:49.814283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tslvs" event={"ID":"99702864-6a6c-4f35-ae70-47a214c4c092","Type":"ContainerStarted","Data":"2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b"} Nov 24 09:06:49 crc kubenswrapper[4799]: I1124 09:06:49.836824 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tslvs" podStartSLOduration=3.152448895 podStartE2EDuration="6.836804595s" podCreationTimestamp="2025-11-24 09:06:43 +0000 UTC" firstStartedPulling="2025-11-24 09:06:45.754741681 +0000 UTC m=+8351.410724155" lastFinishedPulling="2025-11-24 09:06:49.439097371 +0000 UTC m=+8355.095079855" observedRunningTime="2025-11-24 09:06:49.829144857 +0000 UTC m=+8355.485127331" watchObservedRunningTime="2025-11-24 09:06:49.836804595 +0000 UTC m=+8355.492787069" Nov 24 09:06:50 crc kubenswrapper[4799]: I1124 09:06:50.400334 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:06:50 crc kubenswrapper[4799]: I1124 09:06:50.400400 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:06:54 crc kubenswrapper[4799]: I1124 09:06:54.336668 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:54 crc kubenswrapper[4799]: I1124 09:06:54.337035 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:54 crc kubenswrapper[4799]: I1124 09:06:54.389132 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:54 crc kubenswrapper[4799]: I1124 09:06:54.915604 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:54 crc kubenswrapper[4799]: I1124 09:06:54.960538 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tslvs"] Nov 24 09:06:56 crc kubenswrapper[4799]: I1124 09:06:56.884768 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tslvs" podUID="99702864-6a6c-4f35-ae70-47a214c4c092" containerName="registry-server" containerID="cri-o://2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b" gracePeriod=2 Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.351686 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.460278 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-catalog-content\") pod \"99702864-6a6c-4f35-ae70-47a214c4c092\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.460457 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-utilities\") pod \"99702864-6a6c-4f35-ae70-47a214c4c092\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.460640 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvdcp\" (UniqueName: \"kubernetes.io/projected/99702864-6a6c-4f35-ae70-47a214c4c092-kube-api-access-xvdcp\") pod \"99702864-6a6c-4f35-ae70-47a214c4c092\" (UID: \"99702864-6a6c-4f35-ae70-47a214c4c092\") " Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.461358 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-utilities" (OuterVolumeSpecName: "utilities") pod "99702864-6a6c-4f35-ae70-47a214c4c092" (UID: "99702864-6a6c-4f35-ae70-47a214c4c092"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.461907 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.466263 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99702864-6a6c-4f35-ae70-47a214c4c092-kube-api-access-xvdcp" (OuterVolumeSpecName: "kube-api-access-xvdcp") pod "99702864-6a6c-4f35-ae70-47a214c4c092" (UID: "99702864-6a6c-4f35-ae70-47a214c4c092"). InnerVolumeSpecName "kube-api-access-xvdcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.563348 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvdcp\" (UniqueName: \"kubernetes.io/projected/99702864-6a6c-4f35-ae70-47a214c4c092-kube-api-access-xvdcp\") on node \"crc\" DevicePath \"\"" Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.902080 4799 generic.go:334] "Generic (PLEG): container finished" podID="99702864-6a6c-4f35-ae70-47a214c4c092" containerID="2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b" exitCode=0 Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.902154 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tslvs" event={"ID":"99702864-6a6c-4f35-ae70-47a214c4c092","Type":"ContainerDied","Data":"2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b"} Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.902196 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tslvs" event={"ID":"99702864-6a6c-4f35-ae70-47a214c4c092","Type":"ContainerDied","Data":"9a28f6696d6bdeca606aa7c0b22b8142f8e44e32bd6484421071f99366e6af45"} Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.902218 4799 scope.go:117] "RemoveContainer" containerID="2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b" Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.902412 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tslvs" Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.952954 4799 scope.go:117] "RemoveContainer" containerID="280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf" Nov 24 09:06:57 crc kubenswrapper[4799]: I1124 09:06:57.991037 4799 scope.go:117] "RemoveContainer" containerID="726c2410b3df89eba5bcc20f9c0fc472f2595833b53b6818c51dc366823bcf00" Nov 24 09:06:58 crc kubenswrapper[4799]: I1124 09:06:58.019893 4799 scope.go:117] "RemoveContainer" containerID="2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b" Nov 24 09:06:58 crc kubenswrapper[4799]: E1124 09:06:58.020243 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b\": container with ID starting with 2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b not found: ID does not exist" containerID="2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b" Nov 24 09:06:58 crc kubenswrapper[4799]: I1124 09:06:58.020275 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b"} err="failed to get container status \"2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b\": rpc error: code = NotFound desc = could not find container \"2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b\": container with ID starting with 2958b117b53e2b14e6475939bd0fa9deb41db441280c0e1c023e2e344f0a5a3b not found: ID does not exist" Nov 24 09:06:58 crc kubenswrapper[4799]: I1124 09:06:58.020296 4799 scope.go:117] "RemoveContainer" containerID="280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf" Nov 24 09:06:58 crc kubenswrapper[4799]: E1124 09:06:58.020513 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf\": container with ID starting with 280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf not found: ID does not exist" containerID="280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf" Nov 24 09:06:58 crc kubenswrapper[4799]: I1124 09:06:58.020547 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf"} err="failed to get container status \"280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf\": rpc error: code = NotFound desc = could not find container \"280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf\": container with ID starting with 280d40076d9b25552344f3a29e2c93475b989a36ce756c844a916cf578f96fbf not found: ID does not exist" Nov 24 09:06:58 crc kubenswrapper[4799]: I1124 09:06:58.020570 4799 scope.go:117] "RemoveContainer" containerID="726c2410b3df89eba5bcc20f9c0fc472f2595833b53b6818c51dc366823bcf00" Nov 24 09:06:58 crc kubenswrapper[4799]: E1124 09:06:58.021012 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"726c2410b3df89eba5bcc20f9c0fc472f2595833b53b6818c51dc366823bcf00\": container with ID starting with 726c2410b3df89eba5bcc20f9c0fc472f2595833b53b6818c51dc366823bcf00 not found: ID does not exist" containerID="726c2410b3df89eba5bcc20f9c0fc472f2595833b53b6818c51dc366823bcf00" Nov 24 09:06:58 crc kubenswrapper[4799]: I1124 09:06:58.021037 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"726c2410b3df89eba5bcc20f9c0fc472f2595833b53b6818c51dc366823bcf00"} err="failed to get container status \"726c2410b3df89eba5bcc20f9c0fc472f2595833b53b6818c51dc366823bcf00\": rpc error: code = NotFound desc = could not find container \"726c2410b3df89eba5bcc20f9c0fc472f2595833b53b6818c51dc366823bcf00\": container with ID starting with 726c2410b3df89eba5bcc20f9c0fc472f2595833b53b6818c51dc366823bcf00 not found: ID does not exist" Nov 24 09:06:58 crc kubenswrapper[4799]: I1124 09:06:58.499882 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99702864-6a6c-4f35-ae70-47a214c4c092" (UID: "99702864-6a6c-4f35-ae70-47a214c4c092"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:06:58 crc kubenswrapper[4799]: I1124 09:06:58.540429 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tslvs"] Nov 24 09:06:58 crc kubenswrapper[4799]: I1124 09:06:58.557841 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tslvs"] Nov 24 09:06:58 crc kubenswrapper[4799]: I1124 09:06:58.595764 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99702864-6a6c-4f35-ae70-47a214c4c092-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:06:59 crc kubenswrapper[4799]: I1124 09:06:59.640293 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99702864-6a6c-4f35-ae70-47a214c4c092" path="/var/lib/kubelet/pods/99702864-6a6c-4f35-ae70-47a214c4c092/volumes" Nov 24 09:07:12 crc kubenswrapper[4799]: I1124 09:07:12.986569 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-94tzm"] Nov 24 09:07:12 crc kubenswrapper[4799]: E1124 09:07:12.987596 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99702864-6a6c-4f35-ae70-47a214c4c092" containerName="extract-content" Nov 24 09:07:12 crc kubenswrapper[4799]: I1124 09:07:12.987611 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="99702864-6a6c-4f35-ae70-47a214c4c092" containerName="extract-content" Nov 24 09:07:12 crc kubenswrapper[4799]: E1124 09:07:12.987633 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99702864-6a6c-4f35-ae70-47a214c4c092" containerName="extract-utilities" Nov 24 09:07:12 crc kubenswrapper[4799]: I1124 09:07:12.987639 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="99702864-6a6c-4f35-ae70-47a214c4c092" containerName="extract-utilities" Nov 24 09:07:12 crc kubenswrapper[4799]: E1124 09:07:12.987655 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99702864-6a6c-4f35-ae70-47a214c4c092" containerName="registry-server" Nov 24 09:07:12 crc kubenswrapper[4799]: I1124 09:07:12.987661 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="99702864-6a6c-4f35-ae70-47a214c4c092" containerName="registry-server" Nov 24 09:07:12 crc kubenswrapper[4799]: I1124 09:07:12.987910 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="99702864-6a6c-4f35-ae70-47a214c4c092" containerName="registry-server" Nov 24 09:07:12 crc kubenswrapper[4799]: I1124 09:07:12.989654 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.000576 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-94tzm"] Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.012243 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9w2f\" (UniqueName: \"kubernetes.io/projected/aa5d3e70-9436-4b49-99c8-19a63738a785-kube-api-access-z9w2f\") pod \"redhat-marketplace-94tzm\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.012316 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-utilities\") pod \"redhat-marketplace-94tzm\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.012505 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-catalog-content\") pod \"redhat-marketplace-94tzm\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.114485 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-utilities\") pod \"redhat-marketplace-94tzm\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.114656 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-catalog-content\") pod \"redhat-marketplace-94tzm\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.114693 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9w2f\" (UniqueName: \"kubernetes.io/projected/aa5d3e70-9436-4b49-99c8-19a63738a785-kube-api-access-z9w2f\") pod \"redhat-marketplace-94tzm\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.115058 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-utilities\") pod \"redhat-marketplace-94tzm\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.115451 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-catalog-content\") pod \"redhat-marketplace-94tzm\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.142223 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9w2f\" (UniqueName: \"kubernetes.io/projected/aa5d3e70-9436-4b49-99c8-19a63738a785-kube-api-access-z9w2f\") pod \"redhat-marketplace-94tzm\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.311720 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:13 crc kubenswrapper[4799]: I1124 09:07:13.679888 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-94tzm"] Nov 24 09:07:14 crc kubenswrapper[4799]: I1124 09:07:14.053714 4799 generic.go:334] "Generic (PLEG): container finished" podID="aa5d3e70-9436-4b49-99c8-19a63738a785" containerID="f4886444769ba8ad210dec6e2f0e5d96049f9eae1c3efce99033383b616ad36b" exitCode=0 Nov 24 09:07:14 crc kubenswrapper[4799]: I1124 09:07:14.053826 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94tzm" event={"ID":"aa5d3e70-9436-4b49-99c8-19a63738a785","Type":"ContainerDied","Data":"f4886444769ba8ad210dec6e2f0e5d96049f9eae1c3efce99033383b616ad36b"} Nov 24 09:07:14 crc kubenswrapper[4799]: I1124 09:07:14.054173 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94tzm" event={"ID":"aa5d3e70-9436-4b49-99c8-19a63738a785","Type":"ContainerStarted","Data":"b79a46306d14712c7d432fcdad00fdb49da5ccc927dcb6fa898e1b3943182eb4"} Nov 24 09:07:15 crc kubenswrapper[4799]: I1124 09:07:15.065528 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94tzm" event={"ID":"aa5d3e70-9436-4b49-99c8-19a63738a785","Type":"ContainerStarted","Data":"d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393"} Nov 24 09:07:16 crc kubenswrapper[4799]: I1124 09:07:16.077350 4799 generic.go:334] "Generic (PLEG): container finished" podID="aa5d3e70-9436-4b49-99c8-19a63738a785" containerID="d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393" exitCode=0 Nov 24 09:07:16 crc kubenswrapper[4799]: I1124 09:07:16.077388 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94tzm" event={"ID":"aa5d3e70-9436-4b49-99c8-19a63738a785","Type":"ContainerDied","Data":"d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393"} Nov 24 09:07:17 crc kubenswrapper[4799]: I1124 09:07:17.093110 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94tzm" event={"ID":"aa5d3e70-9436-4b49-99c8-19a63738a785","Type":"ContainerStarted","Data":"51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8"} Nov 24 09:07:17 crc kubenswrapper[4799]: I1124 09:07:17.121466 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-94tzm" podStartSLOduration=2.440870531 podStartE2EDuration="5.12145087s" podCreationTimestamp="2025-11-24 09:07:12 +0000 UTC" firstStartedPulling="2025-11-24 09:07:14.056346781 +0000 UTC m=+8379.712329255" lastFinishedPulling="2025-11-24 09:07:16.73692712 +0000 UTC m=+8382.392909594" observedRunningTime="2025-11-24 09:07:17.121384408 +0000 UTC m=+8382.777366892" watchObservedRunningTime="2025-11-24 09:07:17.12145087 +0000 UTC m=+8382.777433344" Nov 24 09:07:20 crc kubenswrapper[4799]: I1124 09:07:20.400192 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:07:20 crc kubenswrapper[4799]: I1124 09:07:20.400522 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:07:23 crc kubenswrapper[4799]: I1124 09:07:23.312549 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:23 crc kubenswrapper[4799]: I1124 09:07:23.313465 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:23 crc kubenswrapper[4799]: I1124 09:07:23.381839 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:24 crc kubenswrapper[4799]: I1124 09:07:24.206868 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:24 crc kubenswrapper[4799]: I1124 09:07:24.265706 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-94tzm"] Nov 24 09:07:26 crc kubenswrapper[4799]: I1124 09:07:26.181608 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-94tzm" podUID="aa5d3e70-9436-4b49-99c8-19a63738a785" containerName="registry-server" containerID="cri-o://51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8" gracePeriod=2 Nov 24 09:07:26 crc kubenswrapper[4799]: I1124 09:07:26.676915 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:26 crc kubenswrapper[4799]: I1124 09:07:26.834204 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-utilities\") pod \"aa5d3e70-9436-4b49-99c8-19a63738a785\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " Nov 24 09:07:26 crc kubenswrapper[4799]: I1124 09:07:26.834279 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9w2f\" (UniqueName: \"kubernetes.io/projected/aa5d3e70-9436-4b49-99c8-19a63738a785-kube-api-access-z9w2f\") pod \"aa5d3e70-9436-4b49-99c8-19a63738a785\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " Nov 24 09:07:26 crc kubenswrapper[4799]: I1124 09:07:26.834357 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-catalog-content\") pod \"aa5d3e70-9436-4b49-99c8-19a63738a785\" (UID: \"aa5d3e70-9436-4b49-99c8-19a63738a785\") " Nov 24 09:07:26 crc kubenswrapper[4799]: I1124 09:07:26.835735 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-utilities" (OuterVolumeSpecName: "utilities") pod "aa5d3e70-9436-4b49-99c8-19a63738a785" (UID: "aa5d3e70-9436-4b49-99c8-19a63738a785"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:07:26 crc kubenswrapper[4799]: I1124 09:07:26.840630 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa5d3e70-9436-4b49-99c8-19a63738a785-kube-api-access-z9w2f" (OuterVolumeSpecName: "kube-api-access-z9w2f") pod "aa5d3e70-9436-4b49-99c8-19a63738a785" (UID: "aa5d3e70-9436-4b49-99c8-19a63738a785"). InnerVolumeSpecName "kube-api-access-z9w2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:07:26 crc kubenswrapper[4799]: I1124 09:07:26.856931 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa5d3e70-9436-4b49-99c8-19a63738a785" (UID: "aa5d3e70-9436-4b49-99c8-19a63738a785"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:07:26 crc kubenswrapper[4799]: I1124 09:07:26.937458 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:26 crc kubenswrapper[4799]: I1124 09:07:26.937499 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9w2f\" (UniqueName: \"kubernetes.io/projected/aa5d3e70-9436-4b49-99c8-19a63738a785-kube-api-access-z9w2f\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:26 crc kubenswrapper[4799]: I1124 09:07:26.937511 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa5d3e70-9436-4b49-99c8-19a63738a785-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.193058 4799 generic.go:334] "Generic (PLEG): container finished" podID="aa5d3e70-9436-4b49-99c8-19a63738a785" containerID="51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8" exitCode=0 Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.194133 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94tzm" event={"ID":"aa5d3e70-9436-4b49-99c8-19a63738a785","Type":"ContainerDied","Data":"51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8"} Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.194233 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94tzm" event={"ID":"aa5d3e70-9436-4b49-99c8-19a63738a785","Type":"ContainerDied","Data":"b79a46306d14712c7d432fcdad00fdb49da5ccc927dcb6fa898e1b3943182eb4"} Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.194306 4799 scope.go:117] "RemoveContainer" containerID="51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8" Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.194501 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94tzm" Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.234249 4799 scope.go:117] "RemoveContainer" containerID="d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393" Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.245038 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-94tzm"] Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.259165 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-94tzm"] Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.303361 4799 scope.go:117] "RemoveContainer" containerID="f4886444769ba8ad210dec6e2f0e5d96049f9eae1c3efce99033383b616ad36b" Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.357273 4799 scope.go:117] "RemoveContainer" containerID="51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8" Nov 24 09:07:27 crc kubenswrapper[4799]: E1124 09:07:27.357786 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8\": container with ID starting with 51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8 not found: ID does not exist" containerID="51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8" Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.357827 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8"} err="failed to get container status \"51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8\": rpc error: code = NotFound desc = could not find container \"51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8\": container with ID starting with 51d97d134965084d5ce8d57654e04cd5c64fabe63f8c0d4eec4c6c847c6a34b8 not found: ID does not exist" Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.357872 4799 scope.go:117] "RemoveContainer" containerID="d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393" Nov 24 09:07:27 crc kubenswrapper[4799]: E1124 09:07:27.358125 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393\": container with ID starting with d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393 not found: ID does not exist" containerID="d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393" Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.358156 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393"} err="failed to get container status \"d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393\": rpc error: code = NotFound desc = could not find container \"d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393\": container with ID starting with d641cf6a58acc6b631c8f0ea41ce0f502f30b9e9674d8fb277211eae5b51e393 not found: ID does not exist" Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.358175 4799 scope.go:117] "RemoveContainer" containerID="f4886444769ba8ad210dec6e2f0e5d96049f9eae1c3efce99033383b616ad36b" Nov 24 09:07:27 crc kubenswrapper[4799]: E1124 09:07:27.358401 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4886444769ba8ad210dec6e2f0e5d96049f9eae1c3efce99033383b616ad36b\": container with ID starting with f4886444769ba8ad210dec6e2f0e5d96049f9eae1c3efce99033383b616ad36b not found: ID does not exist" containerID="f4886444769ba8ad210dec6e2f0e5d96049f9eae1c3efce99033383b616ad36b" Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.358428 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4886444769ba8ad210dec6e2f0e5d96049f9eae1c3efce99033383b616ad36b"} err="failed to get container status \"f4886444769ba8ad210dec6e2f0e5d96049f9eae1c3efce99033383b616ad36b\": rpc error: code = NotFound desc = could not find container \"f4886444769ba8ad210dec6e2f0e5d96049f9eae1c3efce99033383b616ad36b\": container with ID starting with f4886444769ba8ad210dec6e2f0e5d96049f9eae1c3efce99033383b616ad36b not found: ID does not exist" Nov 24 09:07:27 crc kubenswrapper[4799]: I1124 09:07:27.641176 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa5d3e70-9436-4b49-99c8-19a63738a785" path="/var/lib/kubelet/pods/aa5d3e70-9436-4b49-99c8-19a63738a785/volumes" Nov 24 09:07:43 crc kubenswrapper[4799]: I1124 09:07:43.381918 4799 generic.go:334] "Generic (PLEG): container finished" podID="4ad610ec-2cdd-4eea-8d88-18b633d28225" containerID="5ba83a744d8bcc434a455476c2d4a25d27056821d8a7bb28bbfa4e21ddaeb4c3" exitCode=0 Nov 24 09:07:43 crc kubenswrapper[4799]: I1124 09:07:43.382446 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" event={"ID":"4ad610ec-2cdd-4eea-8d88-18b633d28225","Type":"ContainerDied","Data":"5ba83a744d8bcc434a455476c2d4a25d27056821d8a7bb28bbfa4e21ddaeb4c3"} Nov 24 09:07:44 crc kubenswrapper[4799]: I1124 09:07:44.845831 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.005465 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tljwl\" (UniqueName: \"kubernetes.io/projected/4ad610ec-2cdd-4eea-8d88-18b633d28225-kube-api-access-tljwl\") pod \"4ad610ec-2cdd-4eea-8d88-18b633d28225\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.005553 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ceph\") pod \"4ad610ec-2cdd-4eea-8d88-18b633d28225\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.005636 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-1\") pod \"4ad610ec-2cdd-4eea-8d88-18b633d28225\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.005658 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-0\") pod \"4ad610ec-2cdd-4eea-8d88-18b633d28225\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.005697 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-combined-ca-bundle\") pod \"4ad610ec-2cdd-4eea-8d88-18b633d28225\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.005725 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-1\") pod \"4ad610ec-2cdd-4eea-8d88-18b633d28225\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.005838 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-0\") pod \"4ad610ec-2cdd-4eea-8d88-18b633d28225\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.005893 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-0\") pod \"4ad610ec-2cdd-4eea-8d88-18b633d28225\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.005954 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-1\") pod \"4ad610ec-2cdd-4eea-8d88-18b633d28225\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.005994 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ssh-key\") pod \"4ad610ec-2cdd-4eea-8d88-18b633d28225\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.006031 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-inventory\") pod \"4ad610ec-2cdd-4eea-8d88-18b633d28225\" (UID: \"4ad610ec-2cdd-4eea-8d88-18b633d28225\") " Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.012537 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ceph" (OuterVolumeSpecName: "ceph") pod "4ad610ec-2cdd-4eea-8d88-18b633d28225" (UID: "4ad610ec-2cdd-4eea-8d88-18b633d28225"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.014688 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad610ec-2cdd-4eea-8d88-18b633d28225-kube-api-access-tljwl" (OuterVolumeSpecName: "kube-api-access-tljwl") pod "4ad610ec-2cdd-4eea-8d88-18b633d28225" (UID: "4ad610ec-2cdd-4eea-8d88-18b633d28225"). InnerVolumeSpecName "kube-api-access-tljwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.021174 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "4ad610ec-2cdd-4eea-8d88-18b633d28225" (UID: "4ad610ec-2cdd-4eea-8d88-18b633d28225"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.035935 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "4ad610ec-2cdd-4eea-8d88-18b633d28225" (UID: "4ad610ec-2cdd-4eea-8d88-18b633d28225"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.038784 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4ad610ec-2cdd-4eea-8d88-18b633d28225" (UID: "4ad610ec-2cdd-4eea-8d88-18b633d28225"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.040308 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "4ad610ec-2cdd-4eea-8d88-18b633d28225" (UID: "4ad610ec-2cdd-4eea-8d88-18b633d28225"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.040526 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4ad610ec-2cdd-4eea-8d88-18b633d28225" (UID: "4ad610ec-2cdd-4eea-8d88-18b633d28225"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.043073 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4ad610ec-2cdd-4eea-8d88-18b633d28225" (UID: "4ad610ec-2cdd-4eea-8d88-18b633d28225"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.048132 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4ad610ec-2cdd-4eea-8d88-18b633d28225" (UID: "4ad610ec-2cdd-4eea-8d88-18b633d28225"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.053678 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-inventory" (OuterVolumeSpecName: "inventory") pod "4ad610ec-2cdd-4eea-8d88-18b633d28225" (UID: "4ad610ec-2cdd-4eea-8d88-18b633d28225"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.053816 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4ad610ec-2cdd-4eea-8d88-18b633d28225" (UID: "4ad610ec-2cdd-4eea-8d88-18b633d28225"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.108729 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.108772 4799 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.108781 4799 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.108794 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.108806 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.108816 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tljwl\" (UniqueName: \"kubernetes.io/projected/4ad610ec-2cdd-4eea-8d88-18b633d28225-kube-api-access-tljwl\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.108827 4799 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.108839 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.108865 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.108878 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.108889 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4ad610ec-2cdd-4eea-8d88-18b633d28225-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.410660 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" event={"ID":"4ad610ec-2cdd-4eea-8d88-18b633d28225","Type":"ContainerDied","Data":"0119e928a960cf3f43efb224d2914c4a38e80389bbc81f3237a278bcf8a0e9f3"} Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.410714 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0119e928a960cf3f43efb224d2914c4a38e80389bbc81f3237a278bcf8a0e9f3" Nov 24 09:07:45 crc kubenswrapper[4799]: I1124 09:07:45.410769 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26" Nov 24 09:07:50 crc kubenswrapper[4799]: I1124 09:07:50.400530 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:07:50 crc kubenswrapper[4799]: I1124 09:07:50.401158 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:07:50 crc kubenswrapper[4799]: I1124 09:07:50.401203 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 09:07:50 crc kubenswrapper[4799]: I1124 09:07:50.402046 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:07:50 crc kubenswrapper[4799]: I1124 09:07:50.402090 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" gracePeriod=600 Nov 24 09:07:50 crc kubenswrapper[4799]: E1124 09:07:50.521355 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:07:51 crc kubenswrapper[4799]: I1124 09:07:51.530038 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" exitCode=0 Nov 24 09:07:51 crc kubenswrapper[4799]: I1124 09:07:51.530128 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e"} Nov 24 09:07:51 crc kubenswrapper[4799]: I1124 09:07:51.530404 4799 scope.go:117] "RemoveContainer" containerID="6c892fde2c41f62dc151886b156c9c281cd73f67e86c585ace87fe5a19ac868d" Nov 24 09:07:51 crc kubenswrapper[4799]: I1124 09:07:51.531175 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:07:51 crc kubenswrapper[4799]: E1124 09:07:51.531419 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:08:05 crc kubenswrapper[4799]: I1124 09:08:05.640197 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:08:05 crc kubenswrapper[4799]: E1124 09:08:05.641007 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:08:18 crc kubenswrapper[4799]: I1124 09:08:18.628584 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:08:18 crc kubenswrapper[4799]: E1124 09:08:18.629375 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:08:30 crc kubenswrapper[4799]: I1124 09:08:30.629744 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:08:30 crc kubenswrapper[4799]: E1124 09:08:30.631647 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:08:42 crc kubenswrapper[4799]: I1124 09:08:42.629351 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:08:42 crc kubenswrapper[4799]: E1124 09:08:42.630578 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:08:56 crc kubenswrapper[4799]: I1124 09:08:56.628771 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:08:56 crc kubenswrapper[4799]: E1124 09:08:56.629679 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:09:11 crc kubenswrapper[4799]: I1124 09:09:11.628158 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:09:11 crc kubenswrapper[4799]: E1124 09:09:11.628965 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:09:26 crc kubenswrapper[4799]: I1124 09:09:26.628976 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:09:26 crc kubenswrapper[4799]: E1124 09:09:26.629714 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:09:39 crc kubenswrapper[4799]: I1124 09:09:39.630507 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:09:39 crc kubenswrapper[4799]: E1124 09:09:39.631452 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:09:51 crc kubenswrapper[4799]: I1124 09:09:51.628361 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:09:51 crc kubenswrapper[4799]: E1124 09:09:51.629048 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:09:52 crc kubenswrapper[4799]: I1124 09:09:52.476760 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 09:09:52 crc kubenswrapper[4799]: I1124 09:09:52.477428 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="90f89a97-f122-4910-b4be-52bd61722372" containerName="adoption" containerID="cri-o://fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb" gracePeriod=30 Nov 24 09:10:04 crc kubenswrapper[4799]: I1124 09:10:04.629047 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:10:04 crc kubenswrapper[4799]: E1124 09:10:04.630007 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:10:19 crc kubenswrapper[4799]: I1124 09:10:19.629139 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:10:19 crc kubenswrapper[4799]: E1124 09:10:19.630478 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.034266 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.048597 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\") pod \"90f89a97-f122-4910-b4be-52bd61722372\" (UID: \"90f89a97-f122-4910-b4be-52bd61722372\") " Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.049305 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttjgk\" (UniqueName: \"kubernetes.io/projected/90f89a97-f122-4910-b4be-52bd61722372-kube-api-access-ttjgk\") pod \"90f89a97-f122-4910-b4be-52bd61722372\" (UID: \"90f89a97-f122-4910-b4be-52bd61722372\") " Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.059281 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90f89a97-f122-4910-b4be-52bd61722372-kube-api-access-ttjgk" (OuterVolumeSpecName: "kube-api-access-ttjgk") pod "90f89a97-f122-4910-b4be-52bd61722372" (UID: "90f89a97-f122-4910-b4be-52bd61722372"). InnerVolumeSpecName "kube-api-access-ttjgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.070292 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a" (OuterVolumeSpecName: "mariadb-data") pod "90f89a97-f122-4910-b4be-52bd61722372" (UID: "90f89a97-f122-4910-b4be-52bd61722372"). InnerVolumeSpecName "pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.146674 4799 generic.go:334] "Generic (PLEG): container finished" podID="90f89a97-f122-4910-b4be-52bd61722372" containerID="fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb" exitCode=137 Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.146720 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"90f89a97-f122-4910-b4be-52bd61722372","Type":"ContainerDied","Data":"fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb"} Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.146751 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"90f89a97-f122-4910-b4be-52bd61722372","Type":"ContainerDied","Data":"9e07c6b90a00ae2d4041cce3a19a0b2314f5e4b84b3b3848c2fb5060f6bf2061"} Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.146772 4799 scope.go:117] "RemoveContainer" containerID="fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.146940 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.152449 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttjgk\" (UniqueName: \"kubernetes.io/projected/90f89a97-f122-4910-b4be-52bd61722372-kube-api-access-ttjgk\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.152504 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\") on node \"crc\" " Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.178166 4799 scope.go:117] "RemoveContainer" containerID="fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb" Nov 24 09:10:23 crc kubenswrapper[4799]: E1124 09:10:23.178623 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb\": container with ID starting with fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb not found: ID does not exist" containerID="fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.178678 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb"} err="failed to get container status \"fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb\": rpc error: code = NotFound desc = could not find container \"fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb\": container with ID starting with fc6fbdab8c28a86b38d58e94cf10ed96f9d9f9b82ecbda6199c63fd3f13eebcb not found: ID does not exist" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.185355 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.192129 4799 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.192292 4799 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a") on node "crc" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.194517 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.254461 4799 reconciler_common.go:293] "Volume detached for volume \"pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-440cc897-8d6d-4359-a597-0b8f6d1a4b6a\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.642871 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90f89a97-f122-4910-b4be-52bd61722372" path="/var/lib/kubelet/pods/90f89a97-f122-4910-b4be-52bd61722372/volumes" Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.824167 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 09:10:23 crc kubenswrapper[4799]: I1124 09:10:23.824379 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="f7836cca-6a0e-4646-9777-8d706c6cab67" containerName="adoption" containerID="cri-o://19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847" gracePeriod=30 Nov 24 09:10:32 crc kubenswrapper[4799]: I1124 09:10:32.629514 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:10:32 crc kubenswrapper[4799]: E1124 09:10:32.631257 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:10:46 crc kubenswrapper[4799]: I1124 09:10:46.629397 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:10:46 crc kubenswrapper[4799]: E1124 09:10:46.630390 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.325451 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.429109 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/f7836cca-6a0e-4646-9777-8d706c6cab67-ovn-data-cert\") pod \"f7836cca-6a0e-4646-9777-8d706c6cab67\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.430198 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\") pod \"f7836cca-6a0e-4646-9777-8d706c6cab67\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.430363 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f82fc\" (UniqueName: \"kubernetes.io/projected/f7836cca-6a0e-4646-9777-8d706c6cab67-kube-api-access-f82fc\") pod \"f7836cca-6a0e-4646-9777-8d706c6cab67\" (UID: \"f7836cca-6a0e-4646-9777-8d706c6cab67\") " Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.434605 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7836cca-6a0e-4646-9777-8d706c6cab67-kube-api-access-f82fc" (OuterVolumeSpecName: "kube-api-access-f82fc") pod "f7836cca-6a0e-4646-9777-8d706c6cab67" (UID: "f7836cca-6a0e-4646-9777-8d706c6cab67"). InnerVolumeSpecName "kube-api-access-f82fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.435084 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7836cca-6a0e-4646-9777-8d706c6cab67-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "f7836cca-6a0e-4646-9777-8d706c6cab67" (UID: "f7836cca-6a0e-4646-9777-8d706c6cab67"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.454221 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd" (OuterVolumeSpecName: "ovn-data") pod "f7836cca-6a0e-4646-9777-8d706c6cab67" (UID: "f7836cca-6a0e-4646-9777-8d706c6cab67"). InnerVolumeSpecName "pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.494539 4799 generic.go:334] "Generic (PLEG): container finished" podID="f7836cca-6a0e-4646-9777-8d706c6cab67" containerID="19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847" exitCode=137 Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.494582 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"f7836cca-6a0e-4646-9777-8d706c6cab67","Type":"ContainerDied","Data":"19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847"} Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.494609 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"f7836cca-6a0e-4646-9777-8d706c6cab67","Type":"ContainerDied","Data":"8d4b7d62490c3a889a2fcb3d65954e7d01c3e1439d1c7ddd608dd4a6a51ea6d6"} Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.494617 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.494624 4799 scope.go:117] "RemoveContainer" containerID="19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.533135 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\") on node \"crc\" " Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.533171 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f82fc\" (UniqueName: \"kubernetes.io/projected/f7836cca-6a0e-4646-9777-8d706c6cab67-kube-api-access-f82fc\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.533182 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/f7836cca-6a0e-4646-9777-8d706c6cab67-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.547234 4799 scope.go:117] "RemoveContainer" containerID="19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847" Nov 24 09:10:54 crc kubenswrapper[4799]: E1124 09:10:54.547677 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847\": container with ID starting with 19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847 not found: ID does not exist" containerID="19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.547735 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847"} err="failed to get container status \"19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847\": rpc error: code = NotFound desc = could not find container \"19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847\": container with ID starting with 19c97fc449de6c7f59dda3b2113daf786c94090cb9853ac726375ccbf828d847 not found: ID does not exist" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.551690 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.560652 4799 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.560831 4799 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd") on node "crc" Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.562063 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 09:10:54 crc kubenswrapper[4799]: I1124 09:10:54.635222 4799 reconciler_common.go:293] "Volume detached for volume \"pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a0a7fa3-37d8-438f-a64c-e029a7ec05cd\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:55 crc kubenswrapper[4799]: I1124 09:10:55.642453 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7836cca-6a0e-4646-9777-8d706c6cab67" path="/var/lib/kubelet/pods/f7836cca-6a0e-4646-9777-8d706c6cab67/volumes" Nov 24 09:10:57 crc kubenswrapper[4799]: I1124 09:10:57.629090 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:10:57 crc kubenswrapper[4799]: E1124 09:10:57.629674 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:11:12 crc kubenswrapper[4799]: I1124 09:11:12.628728 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:11:12 crc kubenswrapper[4799]: E1124 09:11:12.629429 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.001570 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 09:11:15 crc kubenswrapper[4799]: E1124 09:11:15.002569 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad610ec-2cdd-4eea-8d88-18b633d28225" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.002585 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad610ec-2cdd-4eea-8d88-18b633d28225" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 24 09:11:15 crc kubenswrapper[4799]: E1124 09:11:15.002598 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90f89a97-f122-4910-b4be-52bd61722372" containerName="adoption" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.002604 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="90f89a97-f122-4910-b4be-52bd61722372" containerName="adoption" Nov 24 09:11:15 crc kubenswrapper[4799]: E1124 09:11:15.002622 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5d3e70-9436-4b49-99c8-19a63738a785" containerName="registry-server" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.002627 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5d3e70-9436-4b49-99c8-19a63738a785" containerName="registry-server" Nov 24 09:11:15 crc kubenswrapper[4799]: E1124 09:11:15.002641 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7836cca-6a0e-4646-9777-8d706c6cab67" containerName="adoption" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.002648 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7836cca-6a0e-4646-9777-8d706c6cab67" containerName="adoption" Nov 24 09:11:15 crc kubenswrapper[4799]: E1124 09:11:15.002658 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5d3e70-9436-4b49-99c8-19a63738a785" containerName="extract-content" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.002664 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5d3e70-9436-4b49-99c8-19a63738a785" containerName="extract-content" Nov 24 09:11:15 crc kubenswrapper[4799]: E1124 09:11:15.002678 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5d3e70-9436-4b49-99c8-19a63738a785" containerName="extract-utilities" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.002684 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5d3e70-9436-4b49-99c8-19a63738a785" containerName="extract-utilities" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.002916 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ad610ec-2cdd-4eea-8d88-18b633d28225" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.002933 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="90f89a97-f122-4910-b4be-52bd61722372" containerName="adoption" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.002961 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7836cca-6a0e-4646-9777-8d706c6cab67" containerName="adoption" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.002971 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5d3e70-9436-4b49-99c8-19a63738a785" containerName="registry-server" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.003739 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.005808 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zmjxb" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.006677 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.006963 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.011941 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.018335 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.178270 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.178940 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.179072 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.179200 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.179350 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmb5v\" (UniqueName: \"kubernetes.io/projected/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-kube-api-access-wmb5v\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.179508 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.179638 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.179776 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-config-data\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.179828 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.281818 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.281947 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmb5v\" (UniqueName: \"kubernetes.io/projected/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-kube-api-access-wmb5v\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.281972 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.282006 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.282050 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-config-data\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.282072 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.282113 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.282204 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.282233 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.282724 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.282897 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.283380 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.283474 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-config-data\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.284245 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.290182 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.295873 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.296051 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.299400 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmb5v\" (UniqueName: \"kubernetes.io/projected/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-kube-api-access-wmb5v\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.315077 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " pod="openstack/tempest-tests-tempest" Nov 24 09:11:15 crc kubenswrapper[4799]: I1124 09:11:15.322421 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 09:11:16 crc kubenswrapper[4799]: I1124 09:11:15.951668 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 09:11:16 crc kubenswrapper[4799]: I1124 09:11:15.955266 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:11:16 crc kubenswrapper[4799]: I1124 09:11:16.712426 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8","Type":"ContainerStarted","Data":"43d49e522e6eb554286bb7ebf14d851fe6b6aedcf17725854755c47fe5612fbe"} Nov 24 09:11:27 crc kubenswrapper[4799]: I1124 09:11:27.629973 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:11:27 crc kubenswrapper[4799]: E1124 09:11:27.630896 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:11:38 crc kubenswrapper[4799]: I1124 09:11:38.628750 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:11:38 crc kubenswrapper[4799]: E1124 09:11:38.629558 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:11:50 crc kubenswrapper[4799]: I1124 09:11:50.629667 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:11:50 crc kubenswrapper[4799]: E1124 09:11:50.630962 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:12:03 crc kubenswrapper[4799]: E1124 09:12:03.033449 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:94a34fda2d142cfe9e3097b1d1bd6839" Nov 24 09:12:03 crc kubenswrapper[4799]: E1124 09:12:03.034076 4799 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:94a34fda2d142cfe9e3097b1d1bd6839" Nov 24 09:12:03 crc kubenswrapper[4799]: E1124 09:12:03.034260 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:94a34fda2d142cfe9e3097b1d1bd6839,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wmb5v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(ed0b638f-ace1-45e6-85d4-20ca5ae0ece8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 09:12:03 crc kubenswrapper[4799]: E1124 09:12:03.035553 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" Nov 24 09:12:03 crc kubenswrapper[4799]: E1124 09:12:03.246176 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:94a34fda2d142cfe9e3097b1d1bd6839\\\"\"" pod="openstack/tempest-tests-tempest" podUID="ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" Nov 24 09:12:05 crc kubenswrapper[4799]: I1124 09:12:05.635988 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:12:05 crc kubenswrapper[4799]: E1124 09:12:05.637136 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:12:14 crc kubenswrapper[4799]: I1124 09:12:14.828921 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 09:12:16 crc kubenswrapper[4799]: I1124 09:12:16.374600 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8","Type":"ContainerStarted","Data":"669eeb605a550d149cdfd8de417ef4616cacced6b34c7cb6e1c7e5cacaa64206"} Nov 24 09:12:17 crc kubenswrapper[4799]: I1124 09:12:17.628459 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:12:17 crc kubenswrapper[4799]: E1124 09:12:17.629260 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:12:29 crc kubenswrapper[4799]: I1124 09:12:29.629143 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:12:29 crc kubenswrapper[4799]: E1124 09:12:29.629921 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:12:44 crc kubenswrapper[4799]: I1124 09:12:44.628585 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:12:44 crc kubenswrapper[4799]: E1124 09:12:44.629275 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:12:56 crc kubenswrapper[4799]: I1124 09:12:56.628200 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:12:57 crc kubenswrapper[4799]: I1124 09:12:57.811405 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"9d88578d24b6cf7777086d0a189dede6279badf65705be4b75767e47c765689b"} Nov 24 09:12:57 crc kubenswrapper[4799]: I1124 09:12:57.831694 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=45.96058541 podStartE2EDuration="1m44.831668672s" podCreationTimestamp="2025-11-24 09:11:13 +0000 UTC" firstStartedPulling="2025-11-24 09:11:15.955010073 +0000 UTC m=+8621.610992547" lastFinishedPulling="2025-11-24 09:12:14.826093335 +0000 UTC m=+8680.482075809" observedRunningTime="2025-11-24 09:12:16.397054046 +0000 UTC m=+8682.053036520" watchObservedRunningTime="2025-11-24 09:12:57.831668672 +0000 UTC m=+8723.487651146" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.169694 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t"] Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.172210 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.174589 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.174799 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.202465 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t"] Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.280866 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8e2babdd-bbd5-4767-9d26-187d4536a233-config-volume\") pod \"collect-profiles-29399595-dnq6t\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.280944 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8e2babdd-bbd5-4767-9d26-187d4536a233-secret-volume\") pod \"collect-profiles-29399595-dnq6t\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.281009 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxcxc\" (UniqueName: \"kubernetes.io/projected/8e2babdd-bbd5-4767-9d26-187d4536a233-kube-api-access-bxcxc\") pod \"collect-profiles-29399595-dnq6t\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.382911 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8e2babdd-bbd5-4767-9d26-187d4536a233-config-volume\") pod \"collect-profiles-29399595-dnq6t\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.383008 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8e2babdd-bbd5-4767-9d26-187d4536a233-secret-volume\") pod \"collect-profiles-29399595-dnq6t\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.383102 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxcxc\" (UniqueName: \"kubernetes.io/projected/8e2babdd-bbd5-4767-9d26-187d4536a233-kube-api-access-bxcxc\") pod \"collect-profiles-29399595-dnq6t\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.384321 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8e2babdd-bbd5-4767-9d26-187d4536a233-config-volume\") pod \"collect-profiles-29399595-dnq6t\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.390194 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8e2babdd-bbd5-4767-9d26-187d4536a233-secret-volume\") pod \"collect-profiles-29399595-dnq6t\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.402934 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxcxc\" (UniqueName: \"kubernetes.io/projected/8e2babdd-bbd5-4767-9d26-187d4536a233-kube-api-access-bxcxc\") pod \"collect-profiles-29399595-dnq6t\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:00 crc kubenswrapper[4799]: I1124 09:15:00.502557 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:01 crc kubenswrapper[4799]: I1124 09:15:01.116265 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t"] Nov 24 09:15:01 crc kubenswrapper[4799]: I1124 09:15:01.909300 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2lcbq"] Nov 24 09:15:01 crc kubenswrapper[4799]: I1124 09:15:01.911747 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:01 crc kubenswrapper[4799]: I1124 09:15:01.932540 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2lcbq"] Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.016289 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-catalog-content\") pod \"community-operators-2lcbq\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.016391 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-utilities\") pod \"community-operators-2lcbq\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.016477 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm42b\" (UniqueName: \"kubernetes.io/projected/4c5627e9-3deb-4eed-a3d9-a218a10733d6-kube-api-access-vm42b\") pod \"community-operators-2lcbq\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.118683 4799 generic.go:334] "Generic (PLEG): container finished" podID="8e2babdd-bbd5-4767-9d26-187d4536a233" containerID="4183a0ab86c4e7a5a37c4576bf8d98b4592faf692115daeef3c7fd3ea935e335" exitCode=0 Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.118742 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" event={"ID":"8e2babdd-bbd5-4767-9d26-187d4536a233","Type":"ContainerDied","Data":"4183a0ab86c4e7a5a37c4576bf8d98b4592faf692115daeef3c7fd3ea935e335"} Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.118964 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" event={"ID":"8e2babdd-bbd5-4767-9d26-187d4536a233","Type":"ContainerStarted","Data":"e87548e1a38dba773caa3c526c99200a1ceee2b15922fa487b31e01d072bfc89"} Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.119295 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-catalog-content\") pod \"community-operators-2lcbq\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.119372 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-utilities\") pod \"community-operators-2lcbq\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.119444 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm42b\" (UniqueName: \"kubernetes.io/projected/4c5627e9-3deb-4eed-a3d9-a218a10733d6-kube-api-access-vm42b\") pod \"community-operators-2lcbq\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.119736 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-catalog-content\") pod \"community-operators-2lcbq\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.119756 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-utilities\") pod \"community-operators-2lcbq\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.144084 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm42b\" (UniqueName: \"kubernetes.io/projected/4c5627e9-3deb-4eed-a3d9-a218a10733d6-kube-api-access-vm42b\") pod \"community-operators-2lcbq\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.252086 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:02 crc kubenswrapper[4799]: I1124 09:15:02.945812 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2lcbq"] Nov 24 09:15:03 crc kubenswrapper[4799]: I1124 09:15:03.132984 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lcbq" event={"ID":"4c5627e9-3deb-4eed-a3d9-a218a10733d6","Type":"ContainerStarted","Data":"f6d33652006e34b2e35226e1023dcd9a9fada17a660477d6f2de2c5b8dc34a7e"} Nov 24 09:15:03 crc kubenswrapper[4799]: I1124 09:15:03.783237 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:03 crc kubenswrapper[4799]: I1124 09:15:03.866001 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8e2babdd-bbd5-4767-9d26-187d4536a233-secret-volume\") pod \"8e2babdd-bbd5-4767-9d26-187d4536a233\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " Nov 24 09:15:03 crc kubenswrapper[4799]: I1124 09:15:03.866245 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8e2babdd-bbd5-4767-9d26-187d4536a233-config-volume\") pod \"8e2babdd-bbd5-4767-9d26-187d4536a233\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " Nov 24 09:15:03 crc kubenswrapper[4799]: I1124 09:15:03.866453 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxcxc\" (UniqueName: \"kubernetes.io/projected/8e2babdd-bbd5-4767-9d26-187d4536a233-kube-api-access-bxcxc\") pod \"8e2babdd-bbd5-4767-9d26-187d4536a233\" (UID: \"8e2babdd-bbd5-4767-9d26-187d4536a233\") " Nov 24 09:15:03 crc kubenswrapper[4799]: I1124 09:15:03.867532 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e2babdd-bbd5-4767-9d26-187d4536a233-config-volume" (OuterVolumeSpecName: "config-volume") pod "8e2babdd-bbd5-4767-9d26-187d4536a233" (UID: "8e2babdd-bbd5-4767-9d26-187d4536a233"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:15:03 crc kubenswrapper[4799]: I1124 09:15:03.875233 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e2babdd-bbd5-4767-9d26-187d4536a233-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8e2babdd-bbd5-4767-9d26-187d4536a233" (UID: "8e2babdd-bbd5-4767-9d26-187d4536a233"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:15:03 crc kubenswrapper[4799]: I1124 09:15:03.896579 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e2babdd-bbd5-4767-9d26-187d4536a233-kube-api-access-bxcxc" (OuterVolumeSpecName: "kube-api-access-bxcxc") pod "8e2babdd-bbd5-4767-9d26-187d4536a233" (UID: "8e2babdd-bbd5-4767-9d26-187d4536a233"). InnerVolumeSpecName "kube-api-access-bxcxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:15:03 crc kubenswrapper[4799]: I1124 09:15:03.968682 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxcxc\" (UniqueName: \"kubernetes.io/projected/8e2babdd-bbd5-4767-9d26-187d4536a233-kube-api-access-bxcxc\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:03 crc kubenswrapper[4799]: I1124 09:15:03.969056 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8e2babdd-bbd5-4767-9d26-187d4536a233-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:03 crc kubenswrapper[4799]: I1124 09:15:03.969066 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8e2babdd-bbd5-4767-9d26-187d4536a233-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:04 crc kubenswrapper[4799]: I1124 09:15:04.143800 4799 generic.go:334] "Generic (PLEG): container finished" podID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" containerID="b8b516081c75e7e25048c6969e60e75a04df9e79d5fe7897bb80a110f78804cf" exitCode=0 Nov 24 09:15:04 crc kubenswrapper[4799]: I1124 09:15:04.143898 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lcbq" event={"ID":"4c5627e9-3deb-4eed-a3d9-a218a10733d6","Type":"ContainerDied","Data":"b8b516081c75e7e25048c6969e60e75a04df9e79d5fe7897bb80a110f78804cf"} Nov 24 09:15:04 crc kubenswrapper[4799]: I1124 09:15:04.146904 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" event={"ID":"8e2babdd-bbd5-4767-9d26-187d4536a233","Type":"ContainerDied","Data":"e87548e1a38dba773caa3c526c99200a1ceee2b15922fa487b31e01d072bfc89"} Nov 24 09:15:04 crc kubenswrapper[4799]: I1124 09:15:04.146953 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e87548e1a38dba773caa3c526c99200a1ceee2b15922fa487b31e01d072bfc89" Nov 24 09:15:04 crc kubenswrapper[4799]: I1124 09:15:04.147019 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-dnq6t" Nov 24 09:15:04 crc kubenswrapper[4799]: I1124 09:15:04.865074 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt"] Nov 24 09:15:04 crc kubenswrapper[4799]: I1124 09:15:04.877464 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399550-8vztt"] Nov 24 09:15:05 crc kubenswrapper[4799]: I1124 09:15:05.161389 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lcbq" event={"ID":"4c5627e9-3deb-4eed-a3d9-a218a10733d6","Type":"ContainerStarted","Data":"71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335"} Nov 24 09:15:05 crc kubenswrapper[4799]: I1124 09:15:05.644591 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8321770-f41c-4952-a7d0-a8fa94543a22" path="/var/lib/kubelet/pods/e8321770-f41c-4952-a7d0-a8fa94543a22/volumes" Nov 24 09:15:07 crc kubenswrapper[4799]: I1124 09:15:07.192692 4799 generic.go:334] "Generic (PLEG): container finished" podID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" containerID="71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335" exitCode=0 Nov 24 09:15:07 crc kubenswrapper[4799]: I1124 09:15:07.192975 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lcbq" event={"ID":"4c5627e9-3deb-4eed-a3d9-a218a10733d6","Type":"ContainerDied","Data":"71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335"} Nov 24 09:15:08 crc kubenswrapper[4799]: I1124 09:15:08.204237 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lcbq" event={"ID":"4c5627e9-3deb-4eed-a3d9-a218a10733d6","Type":"ContainerStarted","Data":"622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061"} Nov 24 09:15:12 crc kubenswrapper[4799]: I1124 09:15:12.252446 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:12 crc kubenswrapper[4799]: I1124 09:15:12.253050 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:12 crc kubenswrapper[4799]: I1124 09:15:12.305177 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:12 crc kubenswrapper[4799]: I1124 09:15:12.323466 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2lcbq" podStartSLOduration=7.56621087 podStartE2EDuration="11.323447691s" podCreationTimestamp="2025-11-24 09:15:01 +0000 UTC" firstStartedPulling="2025-11-24 09:15:04.146201762 +0000 UTC m=+8849.802184236" lastFinishedPulling="2025-11-24 09:15:07.903438583 +0000 UTC m=+8853.559421057" observedRunningTime="2025-11-24 09:15:08.234221574 +0000 UTC m=+8853.890204048" watchObservedRunningTime="2025-11-24 09:15:12.323447691 +0000 UTC m=+8857.979430165" Nov 24 09:15:12 crc kubenswrapper[4799]: I1124 09:15:12.531127 4799 scope.go:117] "RemoveContainer" containerID="b64a918bcf5533d6068009fa3af7fdce739558277ccf7a90942505a32f7afe2a" Nov 24 09:15:13 crc kubenswrapper[4799]: I1124 09:15:13.315386 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:13 crc kubenswrapper[4799]: I1124 09:15:13.367514 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2lcbq"] Nov 24 09:15:15 crc kubenswrapper[4799]: I1124 09:15:15.286010 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2lcbq" podUID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" containerName="registry-server" containerID="cri-o://622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061" gracePeriod=2 Nov 24 09:15:15 crc kubenswrapper[4799]: I1124 09:15:15.978151 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.030150 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-utilities\") pod \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.030297 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-catalog-content\") pod \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.030431 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vm42b\" (UniqueName: \"kubernetes.io/projected/4c5627e9-3deb-4eed-a3d9-a218a10733d6-kube-api-access-vm42b\") pod \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\" (UID: \"4c5627e9-3deb-4eed-a3d9-a218a10733d6\") " Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.031706 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-utilities" (OuterVolumeSpecName: "utilities") pod "4c5627e9-3deb-4eed-a3d9-a218a10733d6" (UID: "4c5627e9-3deb-4eed-a3d9-a218a10733d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.037240 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c5627e9-3deb-4eed-a3d9-a218a10733d6-kube-api-access-vm42b" (OuterVolumeSpecName: "kube-api-access-vm42b") pod "4c5627e9-3deb-4eed-a3d9-a218a10733d6" (UID: "4c5627e9-3deb-4eed-a3d9-a218a10733d6"). InnerVolumeSpecName "kube-api-access-vm42b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.085083 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c5627e9-3deb-4eed-a3d9-a218a10733d6" (UID: "4c5627e9-3deb-4eed-a3d9-a218a10733d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.132948 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.132987 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5627e9-3deb-4eed-a3d9-a218a10733d6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.133002 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vm42b\" (UniqueName: \"kubernetes.io/projected/4c5627e9-3deb-4eed-a3d9-a218a10733d6-kube-api-access-vm42b\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.297899 4799 generic.go:334] "Generic (PLEG): container finished" podID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" containerID="622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061" exitCode=0 Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.297962 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lcbq" event={"ID":"4c5627e9-3deb-4eed-a3d9-a218a10733d6","Type":"ContainerDied","Data":"622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061"} Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.297999 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lcbq" event={"ID":"4c5627e9-3deb-4eed-a3d9-a218a10733d6","Type":"ContainerDied","Data":"f6d33652006e34b2e35226e1023dcd9a9fada17a660477d6f2de2c5b8dc34a7e"} Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.298020 4799 scope.go:117] "RemoveContainer" containerID="622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.298197 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lcbq" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.344616 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2lcbq"] Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.355761 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2lcbq"] Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.362604 4799 scope.go:117] "RemoveContainer" containerID="71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.391602 4799 scope.go:117] "RemoveContainer" containerID="b8b516081c75e7e25048c6969e60e75a04df9e79d5fe7897bb80a110f78804cf" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.434392 4799 scope.go:117] "RemoveContainer" containerID="622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061" Nov 24 09:15:16 crc kubenswrapper[4799]: E1124 09:15:16.434817 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061\": container with ID starting with 622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061 not found: ID does not exist" containerID="622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.434869 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061"} err="failed to get container status \"622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061\": rpc error: code = NotFound desc = could not find container \"622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061\": container with ID starting with 622ade856910311e9e71af4dc34b3fdc23bd01ac83b1137f9fb1205d70938061 not found: ID does not exist" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.434897 4799 scope.go:117] "RemoveContainer" containerID="71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335" Nov 24 09:15:16 crc kubenswrapper[4799]: E1124 09:15:16.435198 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335\": container with ID starting with 71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335 not found: ID does not exist" containerID="71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.435225 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335"} err="failed to get container status \"71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335\": rpc error: code = NotFound desc = could not find container \"71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335\": container with ID starting with 71c70ae4d8f82bdbc397c551317e176769ed541c64476f044e7bdb30d27e7335 not found: ID does not exist" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.435243 4799 scope.go:117] "RemoveContainer" containerID="b8b516081c75e7e25048c6969e60e75a04df9e79d5fe7897bb80a110f78804cf" Nov 24 09:15:16 crc kubenswrapper[4799]: E1124 09:15:16.435517 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8b516081c75e7e25048c6969e60e75a04df9e79d5fe7897bb80a110f78804cf\": container with ID starting with b8b516081c75e7e25048c6969e60e75a04df9e79d5fe7897bb80a110f78804cf not found: ID does not exist" containerID="b8b516081c75e7e25048c6969e60e75a04df9e79d5fe7897bb80a110f78804cf" Nov 24 09:15:16 crc kubenswrapper[4799]: I1124 09:15:16.435554 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8b516081c75e7e25048c6969e60e75a04df9e79d5fe7897bb80a110f78804cf"} err="failed to get container status \"b8b516081c75e7e25048c6969e60e75a04df9e79d5fe7897bb80a110f78804cf\": rpc error: code = NotFound desc = could not find container \"b8b516081c75e7e25048c6969e60e75a04df9e79d5fe7897bb80a110f78804cf\": container with ID starting with b8b516081c75e7e25048c6969e60e75a04df9e79d5fe7897bb80a110f78804cf not found: ID does not exist" Nov 24 09:15:17 crc kubenswrapper[4799]: I1124 09:15:17.640146 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" path="/var/lib/kubelet/pods/4c5627e9-3deb-4eed-a3d9-a218a10733d6/volumes" Nov 24 09:15:20 crc kubenswrapper[4799]: I1124 09:15:20.400210 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:15:20 crc kubenswrapper[4799]: I1124 09:15:20.400816 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:15:50 crc kubenswrapper[4799]: I1124 09:15:50.401139 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:15:50 crc kubenswrapper[4799]: I1124 09:15:50.401877 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:16:20 crc kubenswrapper[4799]: I1124 09:16:20.400659 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:16:20 crc kubenswrapper[4799]: I1124 09:16:20.402660 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:16:20 crc kubenswrapper[4799]: I1124 09:16:20.402801 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 09:16:20 crc kubenswrapper[4799]: I1124 09:16:20.403939 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9d88578d24b6cf7777086d0a189dede6279badf65705be4b75767e47c765689b"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:16:20 crc kubenswrapper[4799]: I1124 09:16:20.404124 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://9d88578d24b6cf7777086d0a189dede6279badf65705be4b75767e47c765689b" gracePeriod=600 Nov 24 09:16:20 crc kubenswrapper[4799]: I1124 09:16:20.942460 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="9d88578d24b6cf7777086d0a189dede6279badf65705be4b75767e47c765689b" exitCode=0 Nov 24 09:16:20 crc kubenswrapper[4799]: I1124 09:16:20.942570 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"9d88578d24b6cf7777086d0a189dede6279badf65705be4b75767e47c765689b"} Nov 24 09:16:20 crc kubenswrapper[4799]: I1124 09:16:20.942954 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2"} Nov 24 09:16:20 crc kubenswrapper[4799]: I1124 09:16:20.942993 4799 scope.go:117] "RemoveContainer" containerID="e9a019fbb269970b59f920bb004d448b642388d42437ed046ed5d9f0b2fce70e" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.767269 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlgt"] Nov 24 09:17:54 crc kubenswrapper[4799]: E1124 09:17:54.768607 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" containerName="extract-utilities" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.768623 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" containerName="extract-utilities" Nov 24 09:17:54 crc kubenswrapper[4799]: E1124 09:17:54.768642 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" containerName="registry-server" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.768650 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" containerName="registry-server" Nov 24 09:17:54 crc kubenswrapper[4799]: E1124 09:17:54.768673 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e2babdd-bbd5-4767-9d26-187d4536a233" containerName="collect-profiles" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.768680 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e2babdd-bbd5-4767-9d26-187d4536a233" containerName="collect-profiles" Nov 24 09:17:54 crc kubenswrapper[4799]: E1124 09:17:54.768692 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" containerName="extract-content" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.768699 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" containerName="extract-content" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.768977 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c5627e9-3deb-4eed-a3d9-a218a10733d6" containerName="registry-server" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.769015 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e2babdd-bbd5-4767-9d26-187d4536a233" containerName="collect-profiles" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.770966 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.780256 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlgt"] Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.807426 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-catalog-content\") pod \"redhat-marketplace-8rlgt\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.807655 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9wmk\" (UniqueName: \"kubernetes.io/projected/d3e992cc-93bd-45b6-87f1-f731f50d17bb-kube-api-access-m9wmk\") pod \"redhat-marketplace-8rlgt\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.807788 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-utilities\") pod \"redhat-marketplace-8rlgt\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.908760 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-utilities\") pod \"redhat-marketplace-8rlgt\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.908869 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-catalog-content\") pod \"redhat-marketplace-8rlgt\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.908957 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9wmk\" (UniqueName: \"kubernetes.io/projected/d3e992cc-93bd-45b6-87f1-f731f50d17bb-kube-api-access-m9wmk\") pod \"redhat-marketplace-8rlgt\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.909358 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-utilities\") pod \"redhat-marketplace-8rlgt\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.909540 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-catalog-content\") pod \"redhat-marketplace-8rlgt\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:17:54 crc kubenswrapper[4799]: I1124 09:17:54.934585 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9wmk\" (UniqueName: \"kubernetes.io/projected/d3e992cc-93bd-45b6-87f1-f731f50d17bb-kube-api-access-m9wmk\") pod \"redhat-marketplace-8rlgt\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:17:55 crc kubenswrapper[4799]: I1124 09:17:55.105590 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:17:55 crc kubenswrapper[4799]: I1124 09:17:55.664820 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlgt"] Nov 24 09:17:55 crc kubenswrapper[4799]: I1124 09:17:55.979325 4799 generic.go:334] "Generic (PLEG): container finished" podID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" containerID="328e6c90cbea0b3717738aaf4770bfdf4ddd7986a87b4a5da6b668a004b74742" exitCode=0 Nov 24 09:17:55 crc kubenswrapper[4799]: I1124 09:17:55.979386 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlgt" event={"ID":"d3e992cc-93bd-45b6-87f1-f731f50d17bb","Type":"ContainerDied","Data":"328e6c90cbea0b3717738aaf4770bfdf4ddd7986a87b4a5da6b668a004b74742"} Nov 24 09:17:55 crc kubenswrapper[4799]: I1124 09:17:55.979421 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlgt" event={"ID":"d3e992cc-93bd-45b6-87f1-f731f50d17bb","Type":"ContainerStarted","Data":"b04b093f148f159b350a78bcb139033986b4a5c43946c9ded37734da2517a315"} Nov 24 09:17:55 crc kubenswrapper[4799]: I1124 09:17:55.981202 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:17:58 crc kubenswrapper[4799]: I1124 09:17:58.003338 4799 generic.go:334] "Generic (PLEG): container finished" podID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" containerID="379bb94fa3323e90d345a4ba6dbde380035da73819fbe247df0c7da4a34cd23d" exitCode=0 Nov 24 09:17:58 crc kubenswrapper[4799]: I1124 09:17:58.003418 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlgt" event={"ID":"d3e992cc-93bd-45b6-87f1-f731f50d17bb","Type":"ContainerDied","Data":"379bb94fa3323e90d345a4ba6dbde380035da73819fbe247df0c7da4a34cd23d"} Nov 24 09:17:59 crc kubenswrapper[4799]: I1124 09:17:59.017392 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlgt" event={"ID":"d3e992cc-93bd-45b6-87f1-f731f50d17bb","Type":"ContainerStarted","Data":"3dbe16b7bc14064ccb9ceb6d917040c9fe00ccfeb0b92b4a26f132742117fb96"} Nov 24 09:17:59 crc kubenswrapper[4799]: I1124 09:17:59.041205 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8rlgt" podStartSLOduration=2.615246969 podStartE2EDuration="5.041188129s" podCreationTimestamp="2025-11-24 09:17:54 +0000 UTC" firstStartedPulling="2025-11-24 09:17:55.980988117 +0000 UTC m=+9021.636970591" lastFinishedPulling="2025-11-24 09:17:58.406929277 +0000 UTC m=+9024.062911751" observedRunningTime="2025-11-24 09:17:59.038532613 +0000 UTC m=+9024.694515087" watchObservedRunningTime="2025-11-24 09:17:59.041188129 +0000 UTC m=+9024.697170603" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.036224 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rfvgp"] Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.038736 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.059216 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rfvgp"] Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.172585 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-utilities\") pod \"certified-operators-rfvgp\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.172627 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98zf7\" (UniqueName: \"kubernetes.io/projected/8faec220-40fd-4fe7-866a-c16ef78d4de4-kube-api-access-98zf7\") pod \"certified-operators-rfvgp\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.172718 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-catalog-content\") pod \"certified-operators-rfvgp\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.274929 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-utilities\") pod \"certified-operators-rfvgp\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.274986 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98zf7\" (UniqueName: \"kubernetes.io/projected/8faec220-40fd-4fe7-866a-c16ef78d4de4-kube-api-access-98zf7\") pod \"certified-operators-rfvgp\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.275113 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-catalog-content\") pod \"certified-operators-rfvgp\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.275586 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-utilities\") pod \"certified-operators-rfvgp\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.275674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-catalog-content\") pod \"certified-operators-rfvgp\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.295951 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98zf7\" (UniqueName: \"kubernetes.io/projected/8faec220-40fd-4fe7-866a-c16ef78d4de4-kube-api-access-98zf7\") pod \"certified-operators-rfvgp\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.359727 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:02 crc kubenswrapper[4799]: I1124 09:18:02.894896 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rfvgp"] Nov 24 09:18:03 crc kubenswrapper[4799]: I1124 09:18:03.056091 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfvgp" event={"ID":"8faec220-40fd-4fe7-866a-c16ef78d4de4","Type":"ContainerStarted","Data":"33c28e8e8ae79ea8943b513ec04d8c7646ee0e59ba50861123ceeee89a6333cd"} Nov 24 09:18:04 crc kubenswrapper[4799]: I1124 09:18:04.067388 4799 generic.go:334] "Generic (PLEG): container finished" podID="8faec220-40fd-4fe7-866a-c16ef78d4de4" containerID="0bd29f8f5f0cee848c2e7c738e71c7d7bc77b660bb09f6b32cf3e04118d3debb" exitCode=0 Nov 24 09:18:04 crc kubenswrapper[4799]: I1124 09:18:04.067500 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfvgp" event={"ID":"8faec220-40fd-4fe7-866a-c16ef78d4de4","Type":"ContainerDied","Data":"0bd29f8f5f0cee848c2e7c738e71c7d7bc77b660bb09f6b32cf3e04118d3debb"} Nov 24 09:18:05 crc kubenswrapper[4799]: I1124 09:18:05.106461 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:18:05 crc kubenswrapper[4799]: I1124 09:18:05.106810 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:18:05 crc kubenswrapper[4799]: I1124 09:18:05.170360 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:18:06 crc kubenswrapper[4799]: I1124 09:18:06.090571 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfvgp" event={"ID":"8faec220-40fd-4fe7-866a-c16ef78d4de4","Type":"ContainerStarted","Data":"61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2"} Nov 24 09:18:06 crc kubenswrapper[4799]: I1124 09:18:06.162990 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:18:06 crc kubenswrapper[4799]: I1124 09:18:06.612684 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlgt"] Nov 24 09:18:08 crc kubenswrapper[4799]: I1124 09:18:08.110611 4799 generic.go:334] "Generic (PLEG): container finished" podID="8faec220-40fd-4fe7-866a-c16ef78d4de4" containerID="61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2" exitCode=0 Nov 24 09:18:08 crc kubenswrapper[4799]: I1124 09:18:08.111193 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8rlgt" podUID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" containerName="registry-server" containerID="cri-o://3dbe16b7bc14064ccb9ceb6d917040c9fe00ccfeb0b92b4a26f132742117fb96" gracePeriod=2 Nov 24 09:18:08 crc kubenswrapper[4799]: I1124 09:18:08.110659 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfvgp" event={"ID":"8faec220-40fd-4fe7-866a-c16ef78d4de4","Type":"ContainerDied","Data":"61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2"} Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.122962 4799 generic.go:334] "Generic (PLEG): container finished" podID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" containerID="3dbe16b7bc14064ccb9ceb6d917040c9fe00ccfeb0b92b4a26f132742117fb96" exitCode=0 Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.123052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlgt" event={"ID":"d3e992cc-93bd-45b6-87f1-f731f50d17bb","Type":"ContainerDied","Data":"3dbe16b7bc14064ccb9ceb6d917040c9fe00ccfeb0b92b4a26f132742117fb96"} Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.325082 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.422765 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-catalog-content\") pod \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.422947 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9wmk\" (UniqueName: \"kubernetes.io/projected/d3e992cc-93bd-45b6-87f1-f731f50d17bb-kube-api-access-m9wmk\") pod \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.423001 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-utilities\") pod \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\" (UID: \"d3e992cc-93bd-45b6-87f1-f731f50d17bb\") " Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.427045 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-utilities" (OuterVolumeSpecName: "utilities") pod "d3e992cc-93bd-45b6-87f1-f731f50d17bb" (UID: "d3e992cc-93bd-45b6-87f1-f731f50d17bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.431727 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e992cc-93bd-45b6-87f1-f731f50d17bb-kube-api-access-m9wmk" (OuterVolumeSpecName: "kube-api-access-m9wmk") pod "d3e992cc-93bd-45b6-87f1-f731f50d17bb" (UID: "d3e992cc-93bd-45b6-87f1-f731f50d17bb"). InnerVolumeSpecName "kube-api-access-m9wmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.449038 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3e992cc-93bd-45b6-87f1-f731f50d17bb" (UID: "d3e992cc-93bd-45b6-87f1-f731f50d17bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.526152 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9wmk\" (UniqueName: \"kubernetes.io/projected/d3e992cc-93bd-45b6-87f1-f731f50d17bb-kube-api-access-m9wmk\") on node \"crc\" DevicePath \"\"" Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.526483 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:18:09 crc kubenswrapper[4799]: I1124 09:18:09.526498 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3e992cc-93bd-45b6-87f1-f731f50d17bb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:18:10 crc kubenswrapper[4799]: I1124 09:18:10.135667 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfvgp" event={"ID":"8faec220-40fd-4fe7-866a-c16ef78d4de4","Type":"ContainerStarted","Data":"883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670"} Nov 24 09:18:10 crc kubenswrapper[4799]: I1124 09:18:10.140287 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8rlgt" event={"ID":"d3e992cc-93bd-45b6-87f1-f731f50d17bb","Type":"ContainerDied","Data":"b04b093f148f159b350a78bcb139033986b4a5c43946c9ded37734da2517a315"} Nov 24 09:18:10 crc kubenswrapper[4799]: I1124 09:18:10.140353 4799 scope.go:117] "RemoveContainer" containerID="3dbe16b7bc14064ccb9ceb6d917040c9fe00ccfeb0b92b4a26f132742117fb96" Nov 24 09:18:10 crc kubenswrapper[4799]: I1124 09:18:10.140503 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8rlgt" Nov 24 09:18:10 crc kubenswrapper[4799]: I1124 09:18:10.167434 4799 scope.go:117] "RemoveContainer" containerID="379bb94fa3323e90d345a4ba6dbde380035da73819fbe247df0c7da4a34cd23d" Nov 24 09:18:10 crc kubenswrapper[4799]: I1124 09:18:10.170054 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rfvgp" podStartSLOduration=3.653087279 podStartE2EDuration="8.170035892s" podCreationTimestamp="2025-11-24 09:18:02 +0000 UTC" firstStartedPulling="2025-11-24 09:18:04.069636391 +0000 UTC m=+9029.725618865" lastFinishedPulling="2025-11-24 09:18:08.586585004 +0000 UTC m=+9034.242567478" observedRunningTime="2025-11-24 09:18:10.160445876 +0000 UTC m=+9035.816428350" watchObservedRunningTime="2025-11-24 09:18:10.170035892 +0000 UTC m=+9035.826018356" Nov 24 09:18:10 crc kubenswrapper[4799]: I1124 09:18:10.187465 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlgt"] Nov 24 09:18:10 crc kubenswrapper[4799]: I1124 09:18:10.197896 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8rlgt"] Nov 24 09:18:10 crc kubenswrapper[4799]: I1124 09:18:10.201277 4799 scope.go:117] "RemoveContainer" containerID="328e6c90cbea0b3717738aaf4770bfdf4ddd7986a87b4a5da6b668a004b74742" Nov 24 09:18:11 crc kubenswrapper[4799]: I1124 09:18:11.639685 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" path="/var/lib/kubelet/pods/d3e992cc-93bd-45b6-87f1-f731f50d17bb/volumes" Nov 24 09:18:12 crc kubenswrapper[4799]: I1124 09:18:12.360672 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:12 crc kubenswrapper[4799]: I1124 09:18:12.361075 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:12 crc kubenswrapper[4799]: I1124 09:18:12.411567 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:20 crc kubenswrapper[4799]: I1124 09:18:20.400421 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:18:20 crc kubenswrapper[4799]: I1124 09:18:20.401069 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.016976 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c5dq6"] Nov 24 09:18:22 crc kubenswrapper[4799]: E1124 09:18:22.017765 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" containerName="registry-server" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.017784 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" containerName="registry-server" Nov 24 09:18:22 crc kubenswrapper[4799]: E1124 09:18:22.017831 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" containerName="extract-utilities" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.017840 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" containerName="extract-utilities" Nov 24 09:18:22 crc kubenswrapper[4799]: E1124 09:18:22.017960 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" containerName="extract-content" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.017971 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" containerName="extract-content" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.018192 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e992cc-93bd-45b6-87f1-f731f50d17bb" containerName="registry-server" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.019981 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.047611 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c5dq6"] Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.194322 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lgt4\" (UniqueName: \"kubernetes.io/projected/40eb2d35-dc7b-4e12-860a-e13da529af4c-kube-api-access-6lgt4\") pod \"redhat-operators-c5dq6\" (UID: \"40eb2d35-dc7b-4e12-860a-e13da529af4c\") " pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.194491 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40eb2d35-dc7b-4e12-860a-e13da529af4c-catalog-content\") pod \"redhat-operators-c5dq6\" (UID: \"40eb2d35-dc7b-4e12-860a-e13da529af4c\") " pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.194537 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40eb2d35-dc7b-4e12-860a-e13da529af4c-utilities\") pod \"redhat-operators-c5dq6\" (UID: \"40eb2d35-dc7b-4e12-860a-e13da529af4c\") " pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.296897 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lgt4\" (UniqueName: \"kubernetes.io/projected/40eb2d35-dc7b-4e12-860a-e13da529af4c-kube-api-access-6lgt4\") pod \"redhat-operators-c5dq6\" (UID: \"40eb2d35-dc7b-4e12-860a-e13da529af4c\") " pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.297066 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40eb2d35-dc7b-4e12-860a-e13da529af4c-catalog-content\") pod \"redhat-operators-c5dq6\" (UID: \"40eb2d35-dc7b-4e12-860a-e13da529af4c\") " pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.297116 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40eb2d35-dc7b-4e12-860a-e13da529af4c-utilities\") pod \"redhat-operators-c5dq6\" (UID: \"40eb2d35-dc7b-4e12-860a-e13da529af4c\") " pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.297613 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40eb2d35-dc7b-4e12-860a-e13da529af4c-utilities\") pod \"redhat-operators-c5dq6\" (UID: \"40eb2d35-dc7b-4e12-860a-e13da529af4c\") " pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.297667 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40eb2d35-dc7b-4e12-860a-e13da529af4c-catalog-content\") pod \"redhat-operators-c5dq6\" (UID: \"40eb2d35-dc7b-4e12-860a-e13da529af4c\") " pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.321776 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lgt4\" (UniqueName: \"kubernetes.io/projected/40eb2d35-dc7b-4e12-860a-e13da529af4c-kube-api-access-6lgt4\") pod \"redhat-operators-c5dq6\" (UID: \"40eb2d35-dc7b-4e12-860a-e13da529af4c\") " pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.380171 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.417514 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:22 crc kubenswrapper[4799]: I1124 09:18:22.979835 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c5dq6"] Nov 24 09:18:23 crc kubenswrapper[4799]: I1124 09:18:23.259522 4799 generic.go:334] "Generic (PLEG): container finished" podID="40eb2d35-dc7b-4e12-860a-e13da529af4c" containerID="888bb035afa4524299efd804cd5b23f0511b47cf01d84e485ee216bdeeccdb84" exitCode=0 Nov 24 09:18:23 crc kubenswrapper[4799]: I1124 09:18:23.259568 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c5dq6" event={"ID":"40eb2d35-dc7b-4e12-860a-e13da529af4c","Type":"ContainerDied","Data":"888bb035afa4524299efd804cd5b23f0511b47cf01d84e485ee216bdeeccdb84"} Nov 24 09:18:23 crc kubenswrapper[4799]: I1124 09:18:23.259594 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c5dq6" event={"ID":"40eb2d35-dc7b-4e12-860a-e13da529af4c","Type":"ContainerStarted","Data":"7efb336dee3a05533199f0ce388ac1e3046042308b71836fd8cd9f816da42a49"} Nov 24 09:18:24 crc kubenswrapper[4799]: I1124 09:18:24.609736 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rfvgp"] Nov 24 09:18:24 crc kubenswrapper[4799]: I1124 09:18:24.610377 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rfvgp" podUID="8faec220-40fd-4fe7-866a-c16ef78d4de4" containerName="registry-server" containerID="cri-o://883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670" gracePeriod=2 Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.276041 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.280438 4799 generic.go:334] "Generic (PLEG): container finished" podID="8faec220-40fd-4fe7-866a-c16ef78d4de4" containerID="883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670" exitCode=0 Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.280488 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfvgp" event={"ID":"8faec220-40fd-4fe7-866a-c16ef78d4de4","Type":"ContainerDied","Data":"883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670"} Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.280516 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfvgp" event={"ID":"8faec220-40fd-4fe7-866a-c16ef78d4de4","Type":"ContainerDied","Data":"33c28e8e8ae79ea8943b513ec04d8c7646ee0e59ba50861123ceeee89a6333cd"} Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.280537 4799 scope.go:117] "RemoveContainer" containerID="883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.280513 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rfvgp" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.312235 4799 scope.go:117] "RemoveContainer" containerID="61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.346090 4799 scope.go:117] "RemoveContainer" containerID="0bd29f8f5f0cee848c2e7c738e71c7d7bc77b660bb09f6b32cf3e04118d3debb" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.386796 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98zf7\" (UniqueName: \"kubernetes.io/projected/8faec220-40fd-4fe7-866a-c16ef78d4de4-kube-api-access-98zf7\") pod \"8faec220-40fd-4fe7-866a-c16ef78d4de4\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.386972 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-utilities\") pod \"8faec220-40fd-4fe7-866a-c16ef78d4de4\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.387033 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-catalog-content\") pod \"8faec220-40fd-4fe7-866a-c16ef78d4de4\" (UID: \"8faec220-40fd-4fe7-866a-c16ef78d4de4\") " Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.388217 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-utilities" (OuterVolumeSpecName: "utilities") pod "8faec220-40fd-4fe7-866a-c16ef78d4de4" (UID: "8faec220-40fd-4fe7-866a-c16ef78d4de4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.388567 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.395663 4799 scope.go:117] "RemoveContainer" containerID="883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.396272 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8faec220-40fd-4fe7-866a-c16ef78d4de4-kube-api-access-98zf7" (OuterVolumeSpecName: "kube-api-access-98zf7") pod "8faec220-40fd-4fe7-866a-c16ef78d4de4" (UID: "8faec220-40fd-4fe7-866a-c16ef78d4de4"). InnerVolumeSpecName "kube-api-access-98zf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:18:25 crc kubenswrapper[4799]: E1124 09:18:25.396686 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670\": container with ID starting with 883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670 not found: ID does not exist" containerID="883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.396729 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670"} err="failed to get container status \"883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670\": rpc error: code = NotFound desc = could not find container \"883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670\": container with ID starting with 883f373c06d83373c15f38c5188d0fda679226bda5c5fe8fc9cd36b83e746670 not found: ID does not exist" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.396751 4799 scope.go:117] "RemoveContainer" containerID="61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2" Nov 24 09:18:25 crc kubenswrapper[4799]: E1124 09:18:25.397434 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2\": container with ID starting with 61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2 not found: ID does not exist" containerID="61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.397466 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2"} err="failed to get container status \"61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2\": rpc error: code = NotFound desc = could not find container \"61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2\": container with ID starting with 61bd9522e270697a49dd3d60123c0d8ee8ba260a6829ed01dc0892892d1b82c2 not found: ID does not exist" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.397484 4799 scope.go:117] "RemoveContainer" containerID="0bd29f8f5f0cee848c2e7c738e71c7d7bc77b660bb09f6b32cf3e04118d3debb" Nov 24 09:18:25 crc kubenswrapper[4799]: E1124 09:18:25.397961 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bd29f8f5f0cee848c2e7c738e71c7d7bc77b660bb09f6b32cf3e04118d3debb\": container with ID starting with 0bd29f8f5f0cee848c2e7c738e71c7d7bc77b660bb09f6b32cf3e04118d3debb not found: ID does not exist" containerID="0bd29f8f5f0cee848c2e7c738e71c7d7bc77b660bb09f6b32cf3e04118d3debb" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.398002 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bd29f8f5f0cee848c2e7c738e71c7d7bc77b660bb09f6b32cf3e04118d3debb"} err="failed to get container status \"0bd29f8f5f0cee848c2e7c738e71c7d7bc77b660bb09f6b32cf3e04118d3debb\": rpc error: code = NotFound desc = could not find container \"0bd29f8f5f0cee848c2e7c738e71c7d7bc77b660bb09f6b32cf3e04118d3debb\": container with ID starting with 0bd29f8f5f0cee848c2e7c738e71c7d7bc77b660bb09f6b32cf3e04118d3debb not found: ID does not exist" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.436726 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8faec220-40fd-4fe7-866a-c16ef78d4de4" (UID: "8faec220-40fd-4fe7-866a-c16ef78d4de4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.490365 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98zf7\" (UniqueName: \"kubernetes.io/projected/8faec220-40fd-4fe7-866a-c16ef78d4de4-kube-api-access-98zf7\") on node \"crc\" DevicePath \"\"" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.490411 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8faec220-40fd-4fe7-866a-c16ef78d4de4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.622182 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rfvgp"] Nov 24 09:18:25 crc kubenswrapper[4799]: I1124 09:18:25.643735 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rfvgp"] Nov 24 09:18:27 crc kubenswrapper[4799]: I1124 09:18:27.640763 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8faec220-40fd-4fe7-866a-c16ef78d4de4" path="/var/lib/kubelet/pods/8faec220-40fd-4fe7-866a-c16ef78d4de4/volumes" Nov 24 09:18:35 crc kubenswrapper[4799]: I1124 09:18:35.402214 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c5dq6" event={"ID":"40eb2d35-dc7b-4e12-860a-e13da529af4c","Type":"ContainerStarted","Data":"67a796f1057c8c0b51e38476f1dbd6af6eb7c436296b9b79c6fe622e80b77d41"} Nov 24 09:18:36 crc kubenswrapper[4799]: I1124 09:18:36.414869 4799 generic.go:334] "Generic (PLEG): container finished" podID="40eb2d35-dc7b-4e12-860a-e13da529af4c" containerID="67a796f1057c8c0b51e38476f1dbd6af6eb7c436296b9b79c6fe622e80b77d41" exitCode=0 Nov 24 09:18:36 crc kubenswrapper[4799]: I1124 09:18:36.414921 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c5dq6" event={"ID":"40eb2d35-dc7b-4e12-860a-e13da529af4c","Type":"ContainerDied","Data":"67a796f1057c8c0b51e38476f1dbd6af6eb7c436296b9b79c6fe622e80b77d41"} Nov 24 09:18:38 crc kubenswrapper[4799]: I1124 09:18:38.450491 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c5dq6" event={"ID":"40eb2d35-dc7b-4e12-860a-e13da529af4c","Type":"ContainerStarted","Data":"50feaebb1d5adada16cbc97db9ec3fdd586b51ac5a5e980cd1a0518a3bfbb935"} Nov 24 09:18:38 crc kubenswrapper[4799]: I1124 09:18:38.474714 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c5dq6" podStartSLOduration=3.568668997 podStartE2EDuration="17.474691138s" podCreationTimestamp="2025-11-24 09:18:21 +0000 UTC" firstStartedPulling="2025-11-24 09:18:23.26148168 +0000 UTC m=+9048.917464154" lastFinishedPulling="2025-11-24 09:18:37.167503801 +0000 UTC m=+9062.823486295" observedRunningTime="2025-11-24 09:18:38.469281072 +0000 UTC m=+9064.125263556" watchObservedRunningTime="2025-11-24 09:18:38.474691138 +0000 UTC m=+9064.130673612" Nov 24 09:18:42 crc kubenswrapper[4799]: I1124 09:18:42.380473 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:42 crc kubenswrapper[4799]: I1124 09:18:42.381004 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:43 crc kubenswrapper[4799]: I1124 09:18:43.437046 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c5dq6" podUID="40eb2d35-dc7b-4e12-860a-e13da529af4c" containerName="registry-server" probeResult="failure" output=< Nov 24 09:18:43 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 09:18:43 crc kubenswrapper[4799]: > Nov 24 09:18:50 crc kubenswrapper[4799]: I1124 09:18:50.400775 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:18:50 crc kubenswrapper[4799]: I1124 09:18:50.401389 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:18:52 crc kubenswrapper[4799]: I1124 09:18:52.435576 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:52 crc kubenswrapper[4799]: I1124 09:18:52.501696 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c5dq6" Nov 24 09:18:53 crc kubenswrapper[4799]: I1124 09:18:53.047156 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c5dq6"] Nov 24 09:18:53 crc kubenswrapper[4799]: I1124 09:18:53.226096 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kssmh"] Nov 24 09:18:53 crc kubenswrapper[4799]: I1124 09:18:53.226370 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kssmh" podUID="e3c00960-59e9-4288-a4cd-38fd635552e7" containerName="registry-server" containerID="cri-o://ac4366c346d6c0ba28251c98afde3dfa36b15fba9d2ad3ebfa646406c85f15be" gracePeriod=2 Nov 24 09:18:53 crc kubenswrapper[4799]: I1124 09:18:53.629109 4799 generic.go:334] "Generic (PLEG): container finished" podID="e3c00960-59e9-4288-a4cd-38fd635552e7" containerID="ac4366c346d6c0ba28251c98afde3dfa36b15fba9d2ad3ebfa646406c85f15be" exitCode=0 Nov 24 09:18:53 crc kubenswrapper[4799]: I1124 09:18:53.652352 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kssmh" event={"ID":"e3c00960-59e9-4288-a4cd-38fd635552e7","Type":"ContainerDied","Data":"ac4366c346d6c0ba28251c98afde3dfa36b15fba9d2ad3ebfa646406c85f15be"} Nov 24 09:18:53 crc kubenswrapper[4799]: I1124 09:18:53.779389 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 09:18:53 crc kubenswrapper[4799]: I1124 09:18:53.910071 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24jc9\" (UniqueName: \"kubernetes.io/projected/e3c00960-59e9-4288-a4cd-38fd635552e7-kube-api-access-24jc9\") pod \"e3c00960-59e9-4288-a4cd-38fd635552e7\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " Nov 24 09:18:53 crc kubenswrapper[4799]: I1124 09:18:53.910329 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-utilities\") pod \"e3c00960-59e9-4288-a4cd-38fd635552e7\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " Nov 24 09:18:53 crc kubenswrapper[4799]: I1124 09:18:53.910396 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-catalog-content\") pod \"e3c00960-59e9-4288-a4cd-38fd635552e7\" (UID: \"e3c00960-59e9-4288-a4cd-38fd635552e7\") " Nov 24 09:18:53 crc kubenswrapper[4799]: I1124 09:18:53.911508 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-utilities" (OuterVolumeSpecName: "utilities") pod "e3c00960-59e9-4288-a4cd-38fd635552e7" (UID: "e3c00960-59e9-4288-a4cd-38fd635552e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:18:53 crc kubenswrapper[4799]: I1124 09:18:53.917533 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c00960-59e9-4288-a4cd-38fd635552e7-kube-api-access-24jc9" (OuterVolumeSpecName: "kube-api-access-24jc9") pod "e3c00960-59e9-4288-a4cd-38fd635552e7" (UID: "e3c00960-59e9-4288-a4cd-38fd635552e7"). InnerVolumeSpecName "kube-api-access-24jc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:18:54 crc kubenswrapper[4799]: I1124 09:18:54.012938 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:18:54 crc kubenswrapper[4799]: I1124 09:18:54.012978 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24jc9\" (UniqueName: \"kubernetes.io/projected/e3c00960-59e9-4288-a4cd-38fd635552e7-kube-api-access-24jc9\") on node \"crc\" DevicePath \"\"" Nov 24 09:18:54 crc kubenswrapper[4799]: I1124 09:18:54.047092 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3c00960-59e9-4288-a4cd-38fd635552e7" (UID: "e3c00960-59e9-4288-a4cd-38fd635552e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:18:54 crc kubenswrapper[4799]: I1124 09:18:54.115348 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3c00960-59e9-4288-a4cd-38fd635552e7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:18:54 crc kubenswrapper[4799]: I1124 09:18:54.663457 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kssmh" Nov 24 09:18:54 crc kubenswrapper[4799]: I1124 09:18:54.665769 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kssmh" event={"ID":"e3c00960-59e9-4288-a4cd-38fd635552e7","Type":"ContainerDied","Data":"4fb7cc3bc90a7070680dff8ede81ffc9e14403892606fe751dd5165ec2a84e0e"} Nov 24 09:18:54 crc kubenswrapper[4799]: I1124 09:18:54.665859 4799 scope.go:117] "RemoveContainer" containerID="ac4366c346d6c0ba28251c98afde3dfa36b15fba9d2ad3ebfa646406c85f15be" Nov 24 09:18:54 crc kubenswrapper[4799]: I1124 09:18:54.727571 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kssmh"] Nov 24 09:18:54 crc kubenswrapper[4799]: I1124 09:18:54.738392 4799 scope.go:117] "RemoveContainer" containerID="163d1cd347202e161f77ab1534d2055440d37b6e2a64ab8cd534c1c6c2b2e552" Nov 24 09:18:54 crc kubenswrapper[4799]: I1124 09:18:54.759703 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kssmh"] Nov 24 09:18:54 crc kubenswrapper[4799]: I1124 09:18:54.788318 4799 scope.go:117] "RemoveContainer" containerID="b43d5c9d0d7334aede5264fba202487cb36a0e367ee6ab91ff5b0df6f2d04458" Nov 24 09:18:55 crc kubenswrapper[4799]: I1124 09:18:55.642111 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3c00960-59e9-4288-a4cd-38fd635552e7" path="/var/lib/kubelet/pods/e3c00960-59e9-4288-a4cd-38fd635552e7/volumes" Nov 24 09:19:20 crc kubenswrapper[4799]: I1124 09:19:20.400384 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:19:20 crc kubenswrapper[4799]: I1124 09:19:20.401035 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:19:20 crc kubenswrapper[4799]: I1124 09:19:20.401089 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 09:19:20 crc kubenswrapper[4799]: I1124 09:19:20.402140 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:19:20 crc kubenswrapper[4799]: I1124 09:19:20.402232 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" gracePeriod=600 Nov 24 09:19:20 crc kubenswrapper[4799]: E1124 09:19:20.527188 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:19:20 crc kubenswrapper[4799]: I1124 09:19:20.909609 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" exitCode=0 Nov 24 09:19:20 crc kubenswrapper[4799]: I1124 09:19:20.909672 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2"} Nov 24 09:19:20 crc kubenswrapper[4799]: I1124 09:19:20.910123 4799 scope.go:117] "RemoveContainer" containerID="9d88578d24b6cf7777086d0a189dede6279badf65705be4b75767e47c765689b" Nov 24 09:19:20 crc kubenswrapper[4799]: I1124 09:19:20.913636 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:19:20 crc kubenswrapper[4799]: E1124 09:19:20.914292 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:19:34 crc kubenswrapper[4799]: I1124 09:19:34.627970 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:19:34 crc kubenswrapper[4799]: E1124 09:19:34.628637 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:19:45 crc kubenswrapper[4799]: I1124 09:19:45.635213 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:19:45 crc kubenswrapper[4799]: E1124 09:19:45.636066 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:20:00 crc kubenswrapper[4799]: I1124 09:20:00.628909 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:20:00 crc kubenswrapper[4799]: E1124 09:20:00.629773 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:20:12 crc kubenswrapper[4799]: I1124 09:20:12.628080 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:20:12 crc kubenswrapper[4799]: E1124 09:20:12.628983 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:20:23 crc kubenswrapper[4799]: I1124 09:20:23.628969 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:20:23 crc kubenswrapper[4799]: E1124 09:20:23.629775 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:20:34 crc kubenswrapper[4799]: I1124 09:20:34.629105 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:20:34 crc kubenswrapper[4799]: E1124 09:20:34.629943 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:20:46 crc kubenswrapper[4799]: I1124 09:20:46.627972 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:20:46 crc kubenswrapper[4799]: E1124 09:20:46.628558 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:20:53 crc kubenswrapper[4799]: I1124 09:20:53.086107 4799 generic.go:334] "Generic (PLEG): container finished" podID="ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" containerID="669eeb605a550d149cdfd8de417ef4616cacced6b34c7cb6e1c7e5cacaa64206" exitCode=0 Nov 24 09:20:53 crc kubenswrapper[4799]: I1124 09:20:53.086291 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8","Type":"ContainerDied","Data":"669eeb605a550d149cdfd8de417ef4616cacced6b34c7cb6e1c7e5cacaa64206"} Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.561334 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.665962 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmb5v\" (UniqueName: \"kubernetes.io/projected/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-kube-api-access-wmb5v\") pod \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.666012 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ssh-key\") pod \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.666041 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config\") pod \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.666083 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ca-certs\") pod \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.666114 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.666175 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-config-data\") pod \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.666265 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-workdir\") pod \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.666308 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-temporary\") pod \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.666376 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config-secret\") pod \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\" (UID: \"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8\") " Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.676695 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" (UID: "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.684313 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" (UID: "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.684640 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-config-data" (OuterVolumeSpecName: "config-data") pod "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" (UID: "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.694649 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" (UID: "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.714275 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" (UID: "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.714506 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-kube-api-access-wmb5v" (OuterVolumeSpecName: "kube-api-access-wmb5v") pod "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" (UID: "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8"). InnerVolumeSpecName "kube-api-access-wmb5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.721886 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" (UID: "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.728940 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" (UID: "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.748525 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" (UID: "ed0b638f-ace1-45e6-85d4-20ca5ae0ece8"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.769450 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmb5v\" (UniqueName: \"kubernetes.io/projected/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-kube-api-access-wmb5v\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.769495 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.769509 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.769519 4799 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.769557 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.769570 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.769581 4799 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.769595 4799 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.769607 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ed0b638f-ace1-45e6-85d4-20ca5ae0ece8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.800914 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 24 09:20:54 crc kubenswrapper[4799]: I1124 09:20:54.872122 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:55 crc kubenswrapper[4799]: I1124 09:20:55.107344 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ed0b638f-ace1-45e6-85d4-20ca5ae0ece8","Type":"ContainerDied","Data":"43d49e522e6eb554286bb7ebf14d851fe6b6aedcf17725854755c47fe5612fbe"} Nov 24 09:20:55 crc kubenswrapper[4799]: I1124 09:20:55.107385 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43d49e522e6eb554286bb7ebf14d851fe6b6aedcf17725854755c47fe5612fbe" Nov 24 09:20:55 crc kubenswrapper[4799]: I1124 09:20:55.107476 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.104257 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 09:20:57 crc kubenswrapper[4799]: E1124 09:20:57.105023 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8faec220-40fd-4fe7-866a-c16ef78d4de4" containerName="registry-server" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.105036 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8faec220-40fd-4fe7-866a-c16ef78d4de4" containerName="registry-server" Nov 24 09:20:57 crc kubenswrapper[4799]: E1124 09:20:57.105058 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8faec220-40fd-4fe7-866a-c16ef78d4de4" containerName="extract-content" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.105066 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8faec220-40fd-4fe7-866a-c16ef78d4de4" containerName="extract-content" Nov 24 09:20:57 crc kubenswrapper[4799]: E1124 09:20:57.105085 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8faec220-40fd-4fe7-866a-c16ef78d4de4" containerName="extract-utilities" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.105091 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8faec220-40fd-4fe7-866a-c16ef78d4de4" containerName="extract-utilities" Nov 24 09:20:57 crc kubenswrapper[4799]: E1124 09:20:57.105103 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c00960-59e9-4288-a4cd-38fd635552e7" containerName="extract-content" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.105110 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c00960-59e9-4288-a4cd-38fd635552e7" containerName="extract-content" Nov 24 09:20:57 crc kubenswrapper[4799]: E1124 09:20:57.105122 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c00960-59e9-4288-a4cd-38fd635552e7" containerName="registry-server" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.105128 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c00960-59e9-4288-a4cd-38fd635552e7" containerName="registry-server" Nov 24 09:20:57 crc kubenswrapper[4799]: E1124 09:20:57.105144 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c00960-59e9-4288-a4cd-38fd635552e7" containerName="extract-utilities" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.105150 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c00960-59e9-4288-a4cd-38fd635552e7" containerName="extract-utilities" Nov 24 09:20:57 crc kubenswrapper[4799]: E1124 09:20:57.105162 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" containerName="tempest-tests-tempest-tests-runner" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.105168 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" containerName="tempest-tests-tempest-tests-runner" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.105363 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8faec220-40fd-4fe7-866a-c16ef78d4de4" containerName="registry-server" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.105390 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c00960-59e9-4288-a4cd-38fd635552e7" containerName="registry-server" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.105410 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed0b638f-ace1-45e6-85d4-20ca5ae0ece8" containerName="tempest-tests-tempest-tests-runner" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.106223 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.108777 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zmjxb" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.114872 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.220076 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dc94072d-adf7-497a-bec7-67b42c411edb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.220402 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgvzr\" (UniqueName: \"kubernetes.io/projected/dc94072d-adf7-497a-bec7-67b42c411edb-kube-api-access-fgvzr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dc94072d-adf7-497a-bec7-67b42c411edb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.325020 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dc94072d-adf7-497a-bec7-67b42c411edb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.325412 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dc94072d-adf7-497a-bec7-67b42c411edb\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.326173 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgvzr\" (UniqueName: \"kubernetes.io/projected/dc94072d-adf7-497a-bec7-67b42c411edb-kube-api-access-fgvzr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dc94072d-adf7-497a-bec7-67b42c411edb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.572683 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgvzr\" (UniqueName: \"kubernetes.io/projected/dc94072d-adf7-497a-bec7-67b42c411edb-kube-api-access-fgvzr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dc94072d-adf7-497a-bec7-67b42c411edb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.614080 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dc94072d-adf7-497a-bec7-67b42c411edb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 09:20:57 crc kubenswrapper[4799]: I1124 09:20:57.740035 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 09:20:58 crc kubenswrapper[4799]: I1124 09:20:58.301324 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 09:20:58 crc kubenswrapper[4799]: I1124 09:20:58.628521 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:20:58 crc kubenswrapper[4799]: E1124 09:20:58.628897 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:20:59 crc kubenswrapper[4799]: I1124 09:20:59.149631 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"dc94072d-adf7-497a-bec7-67b42c411edb","Type":"ContainerStarted","Data":"04d70bc012707f625f155cc3049d11fe370d8b3a0b6790d89e0898db73690c2c"} Nov 24 09:21:00 crc kubenswrapper[4799]: I1124 09:21:00.166250 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"dc94072d-adf7-497a-bec7-67b42c411edb","Type":"ContainerStarted","Data":"5b68aa639ec40aeefbc87116871ffa7552211570d34768b2836dd771ee7a8726"} Nov 24 09:21:00 crc kubenswrapper[4799]: I1124 09:21:00.192708 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.746710389 podStartE2EDuration="3.192677148s" podCreationTimestamp="2025-11-24 09:20:57 +0000 UTC" firstStartedPulling="2025-11-24 09:20:58.30750613 +0000 UTC m=+9203.963488604" lastFinishedPulling="2025-11-24 09:20:59.753472889 +0000 UTC m=+9205.409455363" observedRunningTime="2025-11-24 09:21:00.187486889 +0000 UTC m=+9205.843469363" watchObservedRunningTime="2025-11-24 09:21:00.192677148 +0000 UTC m=+9205.848659622" Nov 24 09:21:09 crc kubenswrapper[4799]: I1124 09:21:09.628812 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:21:09 crc kubenswrapper[4799]: E1124 09:21:09.630071 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:21:24 crc kubenswrapper[4799]: I1124 09:21:24.631168 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:21:24 crc kubenswrapper[4799]: E1124 09:21:24.632424 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:21:36 crc kubenswrapper[4799]: I1124 09:21:36.628402 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:21:36 crc kubenswrapper[4799]: E1124 09:21:36.629636 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:21:50 crc kubenswrapper[4799]: I1124 09:21:50.628470 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:21:50 crc kubenswrapper[4799]: E1124 09:21:50.629379 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.043203 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mwth7/must-gather-gs8f2"] Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.047213 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/must-gather-gs8f2" Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.049501 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mwth7"/"openshift-service-ca.crt" Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.049530 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mwth7"/"default-dockercfg-gt78k" Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.052938 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mwth7"/"kube-root-ca.crt" Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.058104 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mwth7/must-gather-gs8f2"] Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.153486 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8bbbf471-6b23-485e-9118-881aec43406c-must-gather-output\") pod \"must-gather-gs8f2\" (UID: \"8bbbf471-6b23-485e-9118-881aec43406c\") " pod="openshift-must-gather-mwth7/must-gather-gs8f2" Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.153897 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-962c4\" (UniqueName: \"kubernetes.io/projected/8bbbf471-6b23-485e-9118-881aec43406c-kube-api-access-962c4\") pod \"must-gather-gs8f2\" (UID: \"8bbbf471-6b23-485e-9118-881aec43406c\") " pod="openshift-must-gather-mwth7/must-gather-gs8f2" Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.255258 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8bbbf471-6b23-485e-9118-881aec43406c-must-gather-output\") pod \"must-gather-gs8f2\" (UID: \"8bbbf471-6b23-485e-9118-881aec43406c\") " pod="openshift-must-gather-mwth7/must-gather-gs8f2" Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.255301 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-962c4\" (UniqueName: \"kubernetes.io/projected/8bbbf471-6b23-485e-9118-881aec43406c-kube-api-access-962c4\") pod \"must-gather-gs8f2\" (UID: \"8bbbf471-6b23-485e-9118-881aec43406c\") " pod="openshift-must-gather-mwth7/must-gather-gs8f2" Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.256066 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8bbbf471-6b23-485e-9118-881aec43406c-must-gather-output\") pod \"must-gather-gs8f2\" (UID: \"8bbbf471-6b23-485e-9118-881aec43406c\") " pod="openshift-must-gather-mwth7/must-gather-gs8f2" Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.588402 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-962c4\" (UniqueName: \"kubernetes.io/projected/8bbbf471-6b23-485e-9118-881aec43406c-kube-api-access-962c4\") pod \"must-gather-gs8f2\" (UID: \"8bbbf471-6b23-485e-9118-881aec43406c\") " pod="openshift-must-gather-mwth7/must-gather-gs8f2" Nov 24 09:21:59 crc kubenswrapper[4799]: I1124 09:21:59.667149 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/must-gather-gs8f2" Nov 24 09:22:00 crc kubenswrapper[4799]: I1124 09:22:00.129450 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mwth7/must-gather-gs8f2"] Nov 24 09:22:00 crc kubenswrapper[4799]: I1124 09:22:00.767976 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mwth7/must-gather-gs8f2" event={"ID":"8bbbf471-6b23-485e-9118-881aec43406c","Type":"ContainerStarted","Data":"d12c8cbdab6101e58a2e7bdfcc48ef4a0a5350d45bf68ae924eec478510fb2ef"} Nov 24 09:22:04 crc kubenswrapper[4799]: I1124 09:22:04.628159 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:22:04 crc kubenswrapper[4799]: E1124 09:22:04.629106 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:22:07 crc kubenswrapper[4799]: I1124 09:22:07.840510 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mwth7/must-gather-gs8f2" event={"ID":"8bbbf471-6b23-485e-9118-881aec43406c","Type":"ContainerStarted","Data":"8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b"} Nov 24 09:22:07 crc kubenswrapper[4799]: I1124 09:22:07.840878 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mwth7/must-gather-gs8f2" event={"ID":"8bbbf471-6b23-485e-9118-881aec43406c","Type":"ContainerStarted","Data":"1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5"} Nov 24 09:22:07 crc kubenswrapper[4799]: I1124 09:22:07.861086 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mwth7/must-gather-gs8f2" podStartSLOduration=2.162296566 podStartE2EDuration="8.861068476s" podCreationTimestamp="2025-11-24 09:21:59 +0000 UTC" firstStartedPulling="2025-11-24 09:22:00.147298968 +0000 UTC m=+9265.803281442" lastFinishedPulling="2025-11-24 09:22:06.846070878 +0000 UTC m=+9272.502053352" observedRunningTime="2025-11-24 09:22:07.852124898 +0000 UTC m=+9273.508107382" watchObservedRunningTime="2025-11-24 09:22:07.861068476 +0000 UTC m=+9273.517050950" Nov 24 09:22:12 crc kubenswrapper[4799]: I1124 09:22:12.520285 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mwth7/crc-debug-b5rk8"] Nov 24 09:22:12 crc kubenswrapper[4799]: I1124 09:22:12.522633 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/crc-debug-b5rk8" Nov 24 09:22:12 crc kubenswrapper[4799]: I1124 09:22:12.687283 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8blkh\" (UniqueName: \"kubernetes.io/projected/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-kube-api-access-8blkh\") pod \"crc-debug-b5rk8\" (UID: \"45d7a3fa-19ea-47d0-8808-20b7d60dab0f\") " pod="openshift-must-gather-mwth7/crc-debug-b5rk8" Nov 24 09:22:12 crc kubenswrapper[4799]: I1124 09:22:12.687477 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-host\") pod \"crc-debug-b5rk8\" (UID: \"45d7a3fa-19ea-47d0-8808-20b7d60dab0f\") " pod="openshift-must-gather-mwth7/crc-debug-b5rk8" Nov 24 09:22:12 crc kubenswrapper[4799]: I1124 09:22:12.789421 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8blkh\" (UniqueName: \"kubernetes.io/projected/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-kube-api-access-8blkh\") pod \"crc-debug-b5rk8\" (UID: \"45d7a3fa-19ea-47d0-8808-20b7d60dab0f\") " pod="openshift-must-gather-mwth7/crc-debug-b5rk8" Nov 24 09:22:12 crc kubenswrapper[4799]: I1124 09:22:12.789871 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-host\") pod \"crc-debug-b5rk8\" (UID: \"45d7a3fa-19ea-47d0-8808-20b7d60dab0f\") " pod="openshift-must-gather-mwth7/crc-debug-b5rk8" Nov 24 09:22:12 crc kubenswrapper[4799]: I1124 09:22:12.789977 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-host\") pod \"crc-debug-b5rk8\" (UID: \"45d7a3fa-19ea-47d0-8808-20b7d60dab0f\") " pod="openshift-must-gather-mwth7/crc-debug-b5rk8" Nov 24 09:22:12 crc kubenswrapper[4799]: I1124 09:22:12.812274 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8blkh\" (UniqueName: \"kubernetes.io/projected/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-kube-api-access-8blkh\") pod \"crc-debug-b5rk8\" (UID: \"45d7a3fa-19ea-47d0-8808-20b7d60dab0f\") " pod="openshift-must-gather-mwth7/crc-debug-b5rk8" Nov 24 09:22:12 crc kubenswrapper[4799]: I1124 09:22:12.843204 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/crc-debug-b5rk8" Nov 24 09:22:12 crc kubenswrapper[4799]: W1124 09:22:12.918814 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45d7a3fa_19ea_47d0_8808_20b7d60dab0f.slice/crio-528abac07e6250cc238da4d569caddc3c9fef4f118b5ad8976476f11541c4330 WatchSource:0}: Error finding container 528abac07e6250cc238da4d569caddc3c9fef4f118b5ad8976476f11541c4330: Status 404 returned error can't find the container with id 528abac07e6250cc238da4d569caddc3c9fef4f118b5ad8976476f11541c4330 Nov 24 09:22:13 crc kubenswrapper[4799]: I1124 09:22:13.901864 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mwth7/crc-debug-b5rk8" event={"ID":"45d7a3fa-19ea-47d0-8808-20b7d60dab0f","Type":"ContainerStarted","Data":"528abac07e6250cc238da4d569caddc3c9fef4f118b5ad8976476f11541c4330"} Nov 24 09:22:16 crc kubenswrapper[4799]: I1124 09:22:16.628464 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:22:16 crc kubenswrapper[4799]: E1124 09:22:16.628975 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:22:27 crc kubenswrapper[4799]: I1124 09:22:27.029412 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mwth7/crc-debug-b5rk8" event={"ID":"45d7a3fa-19ea-47d0-8808-20b7d60dab0f","Type":"ContainerStarted","Data":"dc56773e8e7d5527f6eefb09ef481bb2b3cb9e49108a62d282c703e9ff508613"} Nov 24 09:22:27 crc kubenswrapper[4799]: I1124 09:22:27.057427 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mwth7/crc-debug-b5rk8" podStartSLOduration=1.755673724 podStartE2EDuration="15.057409155s" podCreationTimestamp="2025-11-24 09:22:12 +0000 UTC" firstStartedPulling="2025-11-24 09:22:12.930334593 +0000 UTC m=+9278.586317067" lastFinishedPulling="2025-11-24 09:22:26.232070024 +0000 UTC m=+9291.888052498" observedRunningTime="2025-11-24 09:22:27.050466865 +0000 UTC m=+9292.706449349" watchObservedRunningTime="2025-11-24 09:22:27.057409155 +0000 UTC m=+9292.713391629" Nov 24 09:22:28 crc kubenswrapper[4799]: I1124 09:22:28.628994 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:22:28 crc kubenswrapper[4799]: E1124 09:22:28.629708 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:22:42 crc kubenswrapper[4799]: I1124 09:22:42.629161 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:22:42 crc kubenswrapper[4799]: E1124 09:22:42.630261 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:22:50 crc kubenswrapper[4799]: I1124 09:22:50.256314 4799 generic.go:334] "Generic (PLEG): container finished" podID="45d7a3fa-19ea-47d0-8808-20b7d60dab0f" containerID="dc56773e8e7d5527f6eefb09ef481bb2b3cb9e49108a62d282c703e9ff508613" exitCode=0 Nov 24 09:22:50 crc kubenswrapper[4799]: I1124 09:22:50.256335 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mwth7/crc-debug-b5rk8" event={"ID":"45d7a3fa-19ea-47d0-8808-20b7d60dab0f","Type":"ContainerDied","Data":"dc56773e8e7d5527f6eefb09ef481bb2b3cb9e49108a62d282c703e9ff508613"} Nov 24 09:22:51 crc kubenswrapper[4799]: I1124 09:22:51.388541 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/crc-debug-b5rk8" Nov 24 09:22:51 crc kubenswrapper[4799]: I1124 09:22:51.425028 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mwth7/crc-debug-b5rk8"] Nov 24 09:22:51 crc kubenswrapper[4799]: I1124 09:22:51.437548 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mwth7/crc-debug-b5rk8"] Nov 24 09:22:51 crc kubenswrapper[4799]: I1124 09:22:51.568753 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8blkh\" (UniqueName: \"kubernetes.io/projected/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-kube-api-access-8blkh\") pod \"45d7a3fa-19ea-47d0-8808-20b7d60dab0f\" (UID: \"45d7a3fa-19ea-47d0-8808-20b7d60dab0f\") " Nov 24 09:22:51 crc kubenswrapper[4799]: I1124 09:22:51.568827 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-host\") pod \"45d7a3fa-19ea-47d0-8808-20b7d60dab0f\" (UID: \"45d7a3fa-19ea-47d0-8808-20b7d60dab0f\") " Nov 24 09:22:51 crc kubenswrapper[4799]: I1124 09:22:51.569509 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-host" (OuterVolumeSpecName: "host") pod "45d7a3fa-19ea-47d0-8808-20b7d60dab0f" (UID: "45d7a3fa-19ea-47d0-8808-20b7d60dab0f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:22:51 crc kubenswrapper[4799]: I1124 09:22:51.671950 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-host\") on node \"crc\" DevicePath \"\"" Nov 24 09:22:51 crc kubenswrapper[4799]: I1124 09:22:51.879575 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-kube-api-access-8blkh" (OuterVolumeSpecName: "kube-api-access-8blkh") pod "45d7a3fa-19ea-47d0-8808-20b7d60dab0f" (UID: "45d7a3fa-19ea-47d0-8808-20b7d60dab0f"). InnerVolumeSpecName "kube-api-access-8blkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:22:51 crc kubenswrapper[4799]: I1124 09:22:51.978244 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8blkh\" (UniqueName: \"kubernetes.io/projected/45d7a3fa-19ea-47d0-8808-20b7d60dab0f-kube-api-access-8blkh\") on node \"crc\" DevicePath \"\"" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.281293 4799 scope.go:117] "RemoveContainer" containerID="dc56773e8e7d5527f6eefb09ef481bb2b3cb9e49108a62d282c703e9ff508613" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.281323 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/crc-debug-b5rk8" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.660763 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mwth7/crc-debug-5jbqj"] Nov 24 09:22:52 crc kubenswrapper[4799]: E1124 09:22:52.661648 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d7a3fa-19ea-47d0-8808-20b7d60dab0f" containerName="container-00" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.661668 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d7a3fa-19ea-47d0-8808-20b7d60dab0f" containerName="container-00" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.661949 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d7a3fa-19ea-47d0-8808-20b7d60dab0f" containerName="container-00" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.662869 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/crc-debug-5jbqj" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.792832 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d478f747-b5ad-4946-9e14-4e3be9760791-host\") pod \"crc-debug-5jbqj\" (UID: \"d478f747-b5ad-4946-9e14-4e3be9760791\") " pod="openshift-must-gather-mwth7/crc-debug-5jbqj" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.792948 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb5nt\" (UniqueName: \"kubernetes.io/projected/d478f747-b5ad-4946-9e14-4e3be9760791-kube-api-access-pb5nt\") pod \"crc-debug-5jbqj\" (UID: \"d478f747-b5ad-4946-9e14-4e3be9760791\") " pod="openshift-must-gather-mwth7/crc-debug-5jbqj" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.894890 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d478f747-b5ad-4946-9e14-4e3be9760791-host\") pod \"crc-debug-5jbqj\" (UID: \"d478f747-b5ad-4946-9e14-4e3be9760791\") " pod="openshift-must-gather-mwth7/crc-debug-5jbqj" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.894988 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d478f747-b5ad-4946-9e14-4e3be9760791-host\") pod \"crc-debug-5jbqj\" (UID: \"d478f747-b5ad-4946-9e14-4e3be9760791\") " pod="openshift-must-gather-mwth7/crc-debug-5jbqj" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.895005 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb5nt\" (UniqueName: \"kubernetes.io/projected/d478f747-b5ad-4946-9e14-4e3be9760791-kube-api-access-pb5nt\") pod \"crc-debug-5jbqj\" (UID: \"d478f747-b5ad-4946-9e14-4e3be9760791\") " pod="openshift-must-gather-mwth7/crc-debug-5jbqj" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.918395 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb5nt\" (UniqueName: \"kubernetes.io/projected/d478f747-b5ad-4946-9e14-4e3be9760791-kube-api-access-pb5nt\") pod \"crc-debug-5jbqj\" (UID: \"d478f747-b5ad-4946-9e14-4e3be9760791\") " pod="openshift-must-gather-mwth7/crc-debug-5jbqj" Nov 24 09:22:52 crc kubenswrapper[4799]: I1124 09:22:52.982143 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/crc-debug-5jbqj" Nov 24 09:22:53 crc kubenswrapper[4799]: I1124 09:22:53.293961 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mwth7/crc-debug-5jbqj" event={"ID":"d478f747-b5ad-4946-9e14-4e3be9760791","Type":"ContainerStarted","Data":"422e97de0435ce62ca8e81cf2e0c896060fc9fb9364c705f9e17167cd641fb37"} Nov 24 09:22:53 crc kubenswrapper[4799]: I1124 09:22:53.642837 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45d7a3fa-19ea-47d0-8808-20b7d60dab0f" path="/var/lib/kubelet/pods/45d7a3fa-19ea-47d0-8808-20b7d60dab0f/volumes" Nov 24 09:22:54 crc kubenswrapper[4799]: I1124 09:22:54.303991 4799 generic.go:334] "Generic (PLEG): container finished" podID="d478f747-b5ad-4946-9e14-4e3be9760791" containerID="5e5b970129bf9d7c3acd2f42bfd3c4bae4cd0fa31a825d354d234778a5048ee4" exitCode=1 Nov 24 09:22:54 crc kubenswrapper[4799]: I1124 09:22:54.304084 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mwth7/crc-debug-5jbqj" event={"ID":"d478f747-b5ad-4946-9e14-4e3be9760791","Type":"ContainerDied","Data":"5e5b970129bf9d7c3acd2f42bfd3c4bae4cd0fa31a825d354d234778a5048ee4"} Nov 24 09:22:54 crc kubenswrapper[4799]: I1124 09:22:54.360678 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mwth7/crc-debug-5jbqj"] Nov 24 09:22:54 crc kubenswrapper[4799]: I1124 09:22:54.369219 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mwth7/crc-debug-5jbqj"] Nov 24 09:22:55 crc kubenswrapper[4799]: I1124 09:22:55.426344 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/crc-debug-5jbqj" Nov 24 09:22:55 crc kubenswrapper[4799]: I1124 09:22:55.545675 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb5nt\" (UniqueName: \"kubernetes.io/projected/d478f747-b5ad-4946-9e14-4e3be9760791-kube-api-access-pb5nt\") pod \"d478f747-b5ad-4946-9e14-4e3be9760791\" (UID: \"d478f747-b5ad-4946-9e14-4e3be9760791\") " Nov 24 09:22:55 crc kubenswrapper[4799]: I1124 09:22:55.545811 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d478f747-b5ad-4946-9e14-4e3be9760791-host\") pod \"d478f747-b5ad-4946-9e14-4e3be9760791\" (UID: \"d478f747-b5ad-4946-9e14-4e3be9760791\") " Nov 24 09:22:55 crc kubenswrapper[4799]: I1124 09:22:55.545922 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d478f747-b5ad-4946-9e14-4e3be9760791-host" (OuterVolumeSpecName: "host") pod "d478f747-b5ad-4946-9e14-4e3be9760791" (UID: "d478f747-b5ad-4946-9e14-4e3be9760791"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:22:55 crc kubenswrapper[4799]: I1124 09:22:55.546539 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d478f747-b5ad-4946-9e14-4e3be9760791-host\") on node \"crc\" DevicePath \"\"" Nov 24 09:22:55 crc kubenswrapper[4799]: I1124 09:22:55.558313 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d478f747-b5ad-4946-9e14-4e3be9760791-kube-api-access-pb5nt" (OuterVolumeSpecName: "kube-api-access-pb5nt") pod "d478f747-b5ad-4946-9e14-4e3be9760791" (UID: "d478f747-b5ad-4946-9e14-4e3be9760791"). InnerVolumeSpecName "kube-api-access-pb5nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:22:55 crc kubenswrapper[4799]: I1124 09:22:55.641296 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d478f747-b5ad-4946-9e14-4e3be9760791" path="/var/lib/kubelet/pods/d478f747-b5ad-4946-9e14-4e3be9760791/volumes" Nov 24 09:22:55 crc kubenswrapper[4799]: I1124 09:22:55.649052 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb5nt\" (UniqueName: \"kubernetes.io/projected/d478f747-b5ad-4946-9e14-4e3be9760791-kube-api-access-pb5nt\") on node \"crc\" DevicePath \"\"" Nov 24 09:22:56 crc kubenswrapper[4799]: I1124 09:22:56.327765 4799 scope.go:117] "RemoveContainer" containerID="5e5b970129bf9d7c3acd2f42bfd3c4bae4cd0fa31a825d354d234778a5048ee4" Nov 24 09:22:56 crc kubenswrapper[4799]: I1124 09:22:56.327803 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/crc-debug-5jbqj" Nov 24 09:22:57 crc kubenswrapper[4799]: I1124 09:22:57.629740 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:22:57 crc kubenswrapper[4799]: E1124 09:22:57.629990 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:23:08 crc kubenswrapper[4799]: I1124 09:23:08.628123 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:23:08 crc kubenswrapper[4799]: E1124 09:23:08.628841 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:23:23 crc kubenswrapper[4799]: I1124 09:23:23.630143 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:23:23 crc kubenswrapper[4799]: E1124 09:23:23.631396 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:23:38 crc kubenswrapper[4799]: I1124 09:23:38.629020 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:23:38 crc kubenswrapper[4799]: E1124 09:23:38.629896 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:23:49 crc kubenswrapper[4799]: I1124 09:23:49.629178 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:23:49 crc kubenswrapper[4799]: E1124 09:23:49.630156 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:24:02 crc kubenswrapper[4799]: I1124 09:24:02.629569 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:24:02 crc kubenswrapper[4799]: E1124 09:24:02.630715 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:24:14 crc kubenswrapper[4799]: I1124 09:24:14.628825 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:24:14 crc kubenswrapper[4799]: E1124 09:24:14.629698 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:24:28 crc kubenswrapper[4799]: I1124 09:24:28.628064 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:24:29 crc kubenswrapper[4799]: I1124 09:24:29.205826 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"517c10c0986c1088f2ae888718c16d2c685822643c593b40da162bf2f3673af3"} Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.673642 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6bpms"] Nov 24 09:25:13 crc kubenswrapper[4799]: E1124 09:25:13.674495 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d478f747-b5ad-4946-9e14-4e3be9760791" containerName="container-00" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.674508 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d478f747-b5ad-4946-9e14-4e3be9760791" containerName="container-00" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.674718 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d478f747-b5ad-4946-9e14-4e3be9760791" containerName="container-00" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.676136 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.763815 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6bpms"] Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.853061 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-catalog-content\") pod \"community-operators-6bpms\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.853827 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-utilities\") pod \"community-operators-6bpms\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.853921 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4kf7\" (UniqueName: \"kubernetes.io/projected/eee4b3a4-a03f-425a-9220-fa3c378400b0-kube-api-access-p4kf7\") pod \"community-operators-6bpms\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.955408 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4kf7\" (UniqueName: \"kubernetes.io/projected/eee4b3a4-a03f-425a-9220-fa3c378400b0-kube-api-access-p4kf7\") pod \"community-operators-6bpms\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.955465 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-catalog-content\") pod \"community-operators-6bpms\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.955588 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-utilities\") pod \"community-operators-6bpms\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.956133 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-utilities\") pod \"community-operators-6bpms\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.956321 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-catalog-content\") pod \"community-operators-6bpms\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:13 crc kubenswrapper[4799]: I1124 09:25:13.989052 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4kf7\" (UniqueName: \"kubernetes.io/projected/eee4b3a4-a03f-425a-9220-fa3c378400b0-kube-api-access-p4kf7\") pod \"community-operators-6bpms\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:14 crc kubenswrapper[4799]: I1124 09:25:14.037186 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:14 crc kubenswrapper[4799]: I1124 09:25:14.569047 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6bpms"] Nov 24 09:25:14 crc kubenswrapper[4799]: I1124 09:25:14.642465 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bpms" event={"ID":"eee4b3a4-a03f-425a-9220-fa3c378400b0","Type":"ContainerStarted","Data":"7da08c8cda5a3cd3783e1ad9332f7a151aeecaf3d473aeb6d638db4183bb1f69"} Nov 24 09:25:15 crc kubenswrapper[4799]: I1124 09:25:15.658987 4799 generic.go:334] "Generic (PLEG): container finished" podID="eee4b3a4-a03f-425a-9220-fa3c378400b0" containerID="107ff56d7810e2d6609df92bd7948944fe716cafa6b23de31a80a3911e58ebee" exitCode=0 Nov 24 09:25:15 crc kubenswrapper[4799]: I1124 09:25:15.659273 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bpms" event={"ID":"eee4b3a4-a03f-425a-9220-fa3c378400b0","Type":"ContainerDied","Data":"107ff56d7810e2d6609df92bd7948944fe716cafa6b23de31a80a3911e58ebee"} Nov 24 09:25:15 crc kubenswrapper[4799]: I1124 09:25:15.662481 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:25:16 crc kubenswrapper[4799]: I1124 09:25:16.670556 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bpms" event={"ID":"eee4b3a4-a03f-425a-9220-fa3c378400b0","Type":"ContainerStarted","Data":"9e2b0504beb5d1415f1665564b10ebb4ae219f457a17dd776fbedade1083681c"} Nov 24 09:25:18 crc kubenswrapper[4799]: I1124 09:25:18.694013 4799 generic.go:334] "Generic (PLEG): container finished" podID="eee4b3a4-a03f-425a-9220-fa3c378400b0" containerID="9e2b0504beb5d1415f1665564b10ebb4ae219f457a17dd776fbedade1083681c" exitCode=0 Nov 24 09:25:18 crc kubenswrapper[4799]: I1124 09:25:18.694137 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bpms" event={"ID":"eee4b3a4-a03f-425a-9220-fa3c378400b0","Type":"ContainerDied","Data":"9e2b0504beb5d1415f1665564b10ebb4ae219f457a17dd776fbedade1083681c"} Nov 24 09:25:19 crc kubenswrapper[4799]: I1124 09:25:19.717933 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bpms" event={"ID":"eee4b3a4-a03f-425a-9220-fa3c378400b0","Type":"ContainerStarted","Data":"7450d74b4fd01c18ab37b1e9f505ef18b20bf6c2ff6013257870ed2015f347b2"} Nov 24 09:25:19 crc kubenswrapper[4799]: I1124 09:25:19.745544 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6bpms" podStartSLOduration=3.331644717 podStartE2EDuration="6.745526118s" podCreationTimestamp="2025-11-24 09:25:13 +0000 UTC" firstStartedPulling="2025-11-24 09:25:15.662223034 +0000 UTC m=+9461.318205498" lastFinishedPulling="2025-11-24 09:25:19.076104435 +0000 UTC m=+9464.732086899" observedRunningTime="2025-11-24 09:25:19.734955874 +0000 UTC m=+9465.390938368" watchObservedRunningTime="2025-11-24 09:25:19.745526118 +0000 UTC m=+9465.401508582" Nov 24 09:25:24 crc kubenswrapper[4799]: I1124 09:25:24.037596 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:24 crc kubenswrapper[4799]: I1124 09:25:24.038076 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:24 crc kubenswrapper[4799]: I1124 09:25:24.094660 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:24 crc kubenswrapper[4799]: I1124 09:25:24.826445 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:24 crc kubenswrapper[4799]: I1124 09:25:24.880280 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6bpms"] Nov 24 09:25:26 crc kubenswrapper[4799]: I1124 09:25:26.782676 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6bpms" podUID="eee4b3a4-a03f-425a-9220-fa3c378400b0" containerName="registry-server" containerID="cri-o://7450d74b4fd01c18ab37b1e9f505ef18b20bf6c2ff6013257870ed2015f347b2" gracePeriod=2 Nov 24 09:25:27 crc kubenswrapper[4799]: I1124 09:25:27.796074 4799 generic.go:334] "Generic (PLEG): container finished" podID="eee4b3a4-a03f-425a-9220-fa3c378400b0" containerID="7450d74b4fd01c18ab37b1e9f505ef18b20bf6c2ff6013257870ed2015f347b2" exitCode=0 Nov 24 09:25:27 crc kubenswrapper[4799]: I1124 09:25:27.796117 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bpms" event={"ID":"eee4b3a4-a03f-425a-9220-fa3c378400b0","Type":"ContainerDied","Data":"7450d74b4fd01c18ab37b1e9f505ef18b20bf6c2ff6013257870ed2015f347b2"} Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.212251 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.357618 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4kf7\" (UniqueName: \"kubernetes.io/projected/eee4b3a4-a03f-425a-9220-fa3c378400b0-kube-api-access-p4kf7\") pod \"eee4b3a4-a03f-425a-9220-fa3c378400b0\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.357685 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-catalog-content\") pod \"eee4b3a4-a03f-425a-9220-fa3c378400b0\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.357795 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-utilities\") pod \"eee4b3a4-a03f-425a-9220-fa3c378400b0\" (UID: \"eee4b3a4-a03f-425a-9220-fa3c378400b0\") " Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.359123 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-utilities" (OuterVolumeSpecName: "utilities") pod "eee4b3a4-a03f-425a-9220-fa3c378400b0" (UID: "eee4b3a4-a03f-425a-9220-fa3c378400b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.367170 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eee4b3a4-a03f-425a-9220-fa3c378400b0-kube-api-access-p4kf7" (OuterVolumeSpecName: "kube-api-access-p4kf7") pod "eee4b3a4-a03f-425a-9220-fa3c378400b0" (UID: "eee4b3a4-a03f-425a-9220-fa3c378400b0"). InnerVolumeSpecName "kube-api-access-p4kf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.411153 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eee4b3a4-a03f-425a-9220-fa3c378400b0" (UID: "eee4b3a4-a03f-425a-9220-fa3c378400b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.462339 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4kf7\" (UniqueName: \"kubernetes.io/projected/eee4b3a4-a03f-425a-9220-fa3c378400b0-kube-api-access-p4kf7\") on node \"crc\" DevicePath \"\"" Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.462543 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.462629 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee4b3a4-a03f-425a-9220-fa3c378400b0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.808791 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bpms" event={"ID":"eee4b3a4-a03f-425a-9220-fa3c378400b0","Type":"ContainerDied","Data":"7da08c8cda5a3cd3783e1ad9332f7a151aeecaf3d473aeb6d638db4183bb1f69"} Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.808877 4799 scope.go:117] "RemoveContainer" containerID="7450d74b4fd01c18ab37b1e9f505ef18b20bf6c2ff6013257870ed2015f347b2" Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.808907 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bpms" Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.847187 4799 scope.go:117] "RemoveContainer" containerID="9e2b0504beb5d1415f1665564b10ebb4ae219f457a17dd776fbedade1083681c" Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.855180 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6bpms"] Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.871391 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6bpms"] Nov 24 09:25:28 crc kubenswrapper[4799]: I1124 09:25:28.880624 4799 scope.go:117] "RemoveContainer" containerID="107ff56d7810e2d6609df92bd7948944fe716cafa6b23de31a80a3911e58ebee" Nov 24 09:25:29 crc kubenswrapper[4799]: I1124 09:25:29.643603 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eee4b3a4-a03f-425a-9220-fa3c378400b0" path="/var/lib/kubelet/pods/eee4b3a4-a03f-425a-9220-fa3c378400b0/volumes" Nov 24 09:25:43 crc kubenswrapper[4799]: I1124 09:25:43.716737 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_fc77dc6a-33f2-4d03-a7f8-1699e2d37fab/init-config-reloader/0.log" Nov 24 09:25:44 crc kubenswrapper[4799]: I1124 09:25:44.007766 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_fc77dc6a-33f2-4d03-a7f8-1699e2d37fab/init-config-reloader/0.log" Nov 24 09:25:44 crc kubenswrapper[4799]: I1124 09:25:44.033934 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_fc77dc6a-33f2-4d03-a7f8-1699e2d37fab/alertmanager/0.log" Nov 24 09:25:44 crc kubenswrapper[4799]: I1124 09:25:44.085683 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_fc77dc6a-33f2-4d03-a7f8-1699e2d37fab/config-reloader/0.log" Nov 24 09:25:44 crc kubenswrapper[4799]: I1124 09:25:44.441930 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_fe1b6042-8017-4fd0-b755-76c8d6af4a0b/aodh-api/0.log" Nov 24 09:25:44 crc kubenswrapper[4799]: I1124 09:25:44.463306 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_fe1b6042-8017-4fd0-b755-76c8d6af4a0b/aodh-evaluator/0.log" Nov 24 09:25:44 crc kubenswrapper[4799]: I1124 09:25:44.466051 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_fe1b6042-8017-4fd0-b755-76c8d6af4a0b/aodh-listener/0.log" Nov 24 09:25:44 crc kubenswrapper[4799]: I1124 09:25:44.608667 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_fe1b6042-8017-4fd0-b755-76c8d6af4a0b/aodh-notifier/0.log" Nov 24 09:25:44 crc kubenswrapper[4799]: I1124 09:25:44.733053 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-85f6f78554-znl22_4f2d5e05-6093-4786-8940-c5039d601979/barbican-api/0.log" Nov 24 09:25:44 crc kubenswrapper[4799]: I1124 09:25:44.745674 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-85f6f78554-znl22_4f2d5e05-6093-4786-8940-c5039d601979/barbican-api-log/0.log" Nov 24 09:25:44 crc kubenswrapper[4799]: I1124 09:25:44.924298 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6fb47964bb-kqvtg_54a286e6-dbd2-4423-b15b-6317aec5119c/barbican-keystone-listener/0.log" Nov 24 09:25:45 crc kubenswrapper[4799]: I1124 09:25:45.278300 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-58fd65459-m8dgs_e93b600f-f9b4-4bda-88a0-86713f38ef9e/barbican-worker/0.log" Nov 24 09:25:45 crc kubenswrapper[4799]: I1124 09:25:45.311208 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-58fd65459-m8dgs_e93b600f-f9b4-4bda-88a0-86713f38ef9e/barbican-worker-log/0.log" Nov 24 09:25:45 crc kubenswrapper[4799]: I1124 09:25:45.550074 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6fb47964bb-kqvtg_54a286e6-dbd2-4423-b15b-6317aec5119c/barbican-keystone-listener-log/0.log" Nov 24 09:25:45 crc kubenswrapper[4799]: I1124 09:25:45.550465 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-lkqh5_fefc2996-4fc3-40b2-bf36-82c82ba5039a/bootstrap-openstack-openstack-cell1/0.log" Nov 24 09:25:45 crc kubenswrapper[4799]: I1124 09:25:45.669749 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_06a31c2c-bbbc-4f20-94ce-9575256cf110/ceilometer-central-agent/0.log" Nov 24 09:25:45 crc kubenswrapper[4799]: I1124 09:25:45.776545 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_06a31c2c-bbbc-4f20-94ce-9575256cf110/ceilometer-notification-agent/0.log" Nov 24 09:25:45 crc kubenswrapper[4799]: I1124 09:25:45.814891 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_06a31c2c-bbbc-4f20-94ce-9575256cf110/proxy-httpd/0.log" Nov 24 09:25:45 crc kubenswrapper[4799]: I1124 09:25:45.850380 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_06a31c2c-bbbc-4f20-94ce-9575256cf110/sg-core/0.log" Nov 24 09:25:45 crc kubenswrapper[4799]: I1124 09:25:45.985449 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-7ddgn_2907da3b-efc0-488e-b8bd-6b0e4940a868/ceph-client-openstack-openstack-cell1/0.log" Nov 24 09:25:46 crc kubenswrapper[4799]: I1124 09:25:46.159447 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8eb0d76e-b94e-4120-aee5-395757ddd145/cinder-api/0.log" Nov 24 09:25:46 crc kubenswrapper[4799]: I1124 09:25:46.215325 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8eb0d76e-b94e-4120-aee5-395757ddd145/cinder-api-log/0.log" Nov 24 09:25:46 crc kubenswrapper[4799]: I1124 09:25:46.427096 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_2c91ef74-c8d3-4283-913d-16cd52f50546/probe/0.log" Nov 24 09:25:46 crc kubenswrapper[4799]: I1124 09:25:46.611472 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_2c91ef74-c8d3-4283-913d-16cd52f50546/cinder-backup/0.log" Nov 24 09:25:46 crc kubenswrapper[4799]: I1124 09:25:46.713928 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c14bf25c-677c-493a-a8d5-dec06d445bb8/cinder-scheduler/0.log" Nov 24 09:25:46 crc kubenswrapper[4799]: I1124 09:25:46.950800 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c14bf25c-677c-493a-a8d5-dec06d445bb8/probe/0.log" Nov 24 09:25:47 crc kubenswrapper[4799]: I1124 09:25:47.013243 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_b55c4e6c-8c8b-4623-b420-3cfb9eea5764/cinder-volume/0.log" Nov 24 09:25:47 crc kubenswrapper[4799]: I1124 09:25:47.066327 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_b55c4e6c-8c8b-4623-b420-3cfb9eea5764/probe/0.log" Nov 24 09:25:47 crc kubenswrapper[4799]: I1124 09:25:47.261689 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-99xmg_d3b5e760-ee75-4fe6-8f12-f8cde96523c9/configure-network-openstack-openstack-cell1/0.log" Nov 24 09:25:47 crc kubenswrapper[4799]: I1124 09:25:47.502559 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-r7zn5_c29d6dd7-fd78-49e1-afcb-7b8349f6fe22/configure-os-openstack-openstack-cell1/0.log" Nov 24 09:25:47 crc kubenswrapper[4799]: I1124 09:25:47.616085 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79466848d7-cb4rm_ac9ec906-a164-43a8-8d5c-274875acb978/init/0.log" Nov 24 09:25:47 crc kubenswrapper[4799]: I1124 09:25:47.773552 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79466848d7-cb4rm_ac9ec906-a164-43a8-8d5c-274875acb978/init/0.log" Nov 24 09:25:47 crc kubenswrapper[4799]: I1124 09:25:47.838660 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79466848d7-cb4rm_ac9ec906-a164-43a8-8d5c-274875acb978/dnsmasq-dns/0.log" Nov 24 09:25:47 crc kubenswrapper[4799]: I1124 09:25:47.930965 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-blvrs_77211556-d2b3-4243-8518-2bfbe5ef543a/download-cache-openstack-openstack-cell1/0.log" Nov 24 09:25:48 crc kubenswrapper[4799]: I1124 09:25:48.542737 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ea055bb9-dd38-48ef-a248-83b68e3176b6/glance-httpd/0.log" Nov 24 09:25:48 crc kubenswrapper[4799]: I1124 09:25:48.563664 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ea055bb9-dd38-48ef-a248-83b68e3176b6/glance-log/0.log" Nov 24 09:25:48 crc kubenswrapper[4799]: I1124 09:25:48.604476 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7f4a7265-94b0-4631-9500-4ca4e8e4af26/glance-log/0.log" Nov 24 09:25:48 crc kubenswrapper[4799]: I1124 09:25:48.659108 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7f4a7265-94b0-4631-9500-4ca4e8e4af26/glance-httpd/0.log" Nov 24 09:25:48 crc kubenswrapper[4799]: I1124 09:25:48.950699 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-57cccd9dc4-tq5ch_5fbab1f1-fb2a-4d63-9f7f-196b57104f8b/heat-api/0.log" Nov 24 09:25:48 crc kubenswrapper[4799]: I1124 09:25:48.999455 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-65f685584f-5fhns_f02d4c9b-ef9f-4109-ae64-0fe7e07e321c/heat-cfnapi/0.log" Nov 24 09:25:49 crc kubenswrapper[4799]: I1124 09:25:49.052444 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-6ddc665688-m6glq_188c6516-aca5-4f29-80da-3921fbb7728f/heat-engine/0.log" Nov 24 09:25:49 crc kubenswrapper[4799]: I1124 09:25:49.203097 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7cd9788b67-pld4l_75ed23b8-89f0-4527-bff2-5c8e60757c8f/horizon/0.log" Nov 24 09:25:49 crc kubenswrapper[4799]: I1124 09:25:49.291159 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7cd9788b67-pld4l_75ed23b8-89f0-4527-bff2-5c8e60757c8f/horizon-log/0.log" Nov 24 09:25:49 crc kubenswrapper[4799]: I1124 09:25:49.517431 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-wzx69_b6c18c61-453f-4e00-ad20-6152e65992e7/install-certs-openstack-openstack-cell1/0.log" Nov 24 09:25:49 crc kubenswrapper[4799]: I1124 09:25:49.594425 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-9nqzd_b80d7525-1ef0-4824-857a-5264a31f9b73/install-os-openstack-openstack-cell1/0.log" Nov 24 09:25:49 crc kubenswrapper[4799]: I1124 09:25:49.804282 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29399581-8scpq_adc8c7bd-9458-4094-8562-65691c517956/keystone-cron/0.log" Nov 24 09:25:50 crc kubenswrapper[4799]: I1124 09:25:50.240868 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-658ddd8bf4-qjlrl_a963eddd-190c-4a8b-adda-9971855f9beb/keystone-api/0.log" Nov 24 09:25:50 crc kubenswrapper[4799]: I1124 09:25:50.308945 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_051a8eb1-f406-4ab1-b59b-30754d729d0e/kube-state-metrics/0.log" Nov 24 09:25:50 crc kubenswrapper[4799]: I1124 09:25:50.358991 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-nnrmm_2ebcd170-a63a-4355-9046-17e8f42f4dc8/libvirt-openstack-openstack-cell1/0.log" Nov 24 09:25:50 crc kubenswrapper[4799]: I1124 09:25:50.592681 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_0510e3b8-af09-4b89-8041-a5810b6fdad7/manila-api-log/0.log" Nov 24 09:25:50 crc kubenswrapper[4799]: I1124 09:25:50.676003 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_99a84edf-0397-4af1-9a2f-252678c9560e/manila-scheduler/0.log" Nov 24 09:25:50 crc kubenswrapper[4799]: I1124 09:25:50.713770 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_0510e3b8-af09-4b89-8041-a5810b6fdad7/manila-api/0.log" Nov 24 09:25:50 crc kubenswrapper[4799]: I1124 09:25:50.736296 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_99a84edf-0397-4af1-9a2f-252678c9560e/probe/0.log" Nov 24 09:25:50 crc kubenswrapper[4799]: I1124 09:25:50.870578 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_02d972f3-0f96-40d9-94e0-949b605be376/probe/0.log" Nov 24 09:25:50 crc kubenswrapper[4799]: I1124 09:25:50.950565 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_02d972f3-0f96-40d9-94e0-949b605be376/manila-share/0.log" Nov 24 09:25:51 crc kubenswrapper[4799]: I1124 09:25:51.463572 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-8684b7c74f-6msxl_74eeb824-c788-48c1-b0d6-0d230cef156f/neutron-httpd/0.log" Nov 24 09:25:51 crc kubenswrapper[4799]: I1124 09:25:51.464503 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-hnqs2_4928abdf-af58-44e6-8469-b6dc46f76e93/neutron-dhcp-openstack-openstack-cell1/0.log" Nov 24 09:25:51 crc kubenswrapper[4799]: I1124 09:25:51.726955 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-8684b7c74f-6msxl_74eeb824-c788-48c1-b0d6-0d230cef156f/neutron-api/0.log" Nov 24 09:25:51 crc kubenswrapper[4799]: I1124 09:25:51.758270 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-ts42m_4b9803e3-b71b-4cb3-a8ec-098180db8e5c/neutron-metadata-openstack-openstack-cell1/0.log" Nov 24 09:25:51 crc kubenswrapper[4799]: I1124 09:25:51.858613 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-svg68_366926c4-79c8-47ec-be98-467421650c86/neutron-sriov-openstack-openstack-cell1/0.log" Nov 24 09:25:52 crc kubenswrapper[4799]: I1124 09:25:52.217176 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0863a6f5-d8a7-4c50-a33b-3398e262a9e7/nova-api-api/0.log" Nov 24 09:25:52 crc kubenswrapper[4799]: I1124 09:25:52.408720 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0863a6f5-d8a7-4c50-a33b-3398e262a9e7/nova-api-log/0.log" Nov 24 09:25:52 crc kubenswrapper[4799]: I1124 09:25:52.426537 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_4ce8ea8c-013d-4f6a-b1a5-11d1d44c67f6/nova-cell0-conductor-conductor/0.log" Nov 24 09:25:52 crc kubenswrapper[4799]: I1124 09:25:52.553902 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_8e3b0342-d46b-42ef-90f8-c46be0e942b8/nova-cell1-conductor-conductor/0.log" Nov 24 09:25:52 crc kubenswrapper[4799]: I1124 09:25:52.776354 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_fc1ba947-a1ac-46e3-91a0-f47ead0e6add/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 09:25:52 crc kubenswrapper[4799]: I1124 09:25:52.939819 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4z26_4ad610ec-2cdd-4eea-8d88-18b633d28225/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Nov 24 09:25:53 crc kubenswrapper[4799]: I1124 09:25:53.199228 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-tfzxs_9356d471-b239-437f-8db4-f9be28f1c3dd/nova-cell1-openstack-openstack-cell1/0.log" Nov 24 09:25:53 crc kubenswrapper[4799]: I1124 09:25:53.304587 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_2238727c-e84a-4641-b168-6168ec24a46a/nova-metadata-log/0.log" Nov 24 09:25:53 crc kubenswrapper[4799]: I1124 09:25:53.418539 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_2238727c-e84a-4641-b168-6168ec24a46a/nova-metadata-metadata/0.log" Nov 24 09:25:53 crc kubenswrapper[4799]: I1124 09:25:53.616528 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_17aeb60b-35e6-4096-bdce-26ac7c14797c/nova-scheduler-scheduler/0.log" Nov 24 09:25:53 crc kubenswrapper[4799]: I1124 09:25:53.700044 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d77e145-8599-45f3-ab56-13e97d34183e/mysql-bootstrap/0.log" Nov 24 09:25:53 crc kubenswrapper[4799]: I1124 09:25:53.887747 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d77e145-8599-45f3-ab56-13e97d34183e/mysql-bootstrap/0.log" Nov 24 09:25:53 crc kubenswrapper[4799]: I1124 09:25:53.968662 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2d77e145-8599-45f3-ab56-13e97d34183e/galera/0.log" Nov 24 09:25:53 crc kubenswrapper[4799]: I1124 09:25:53.989776 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_63e8e2e8-1c56-48b1-a0fe-5626a960209f/mysql-bootstrap/0.log" Nov 24 09:25:54 crc kubenswrapper[4799]: I1124 09:25:54.187939 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_63e8e2e8-1c56-48b1-a0fe-5626a960209f/mysql-bootstrap/0.log" Nov 24 09:25:54 crc kubenswrapper[4799]: I1124 09:25:54.260195 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_63e8e2e8-1c56-48b1-a0fe-5626a960209f/galera/0.log" Nov 24 09:25:54 crc kubenswrapper[4799]: I1124 09:25:54.269363 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2093fc37-b94e-4a73-b661-d73b3f121c68/openstackclient/0.log" Nov 24 09:25:54 crc kubenswrapper[4799]: I1124 09:25:54.499283 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e7351300-d244-4acd-a82d-a2696a724d92/openstack-network-exporter/0.log" Nov 24 09:25:54 crc kubenswrapper[4799]: I1124 09:25:54.531318 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e7351300-d244-4acd-a82d-a2696a724d92/ovn-northd/0.log" Nov 24 09:25:54 crc kubenswrapper[4799]: I1124 09:25:54.774965 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2b5580f4-5a36-492a-a9de-cdae05bda4a7/openstack-network-exporter/0.log" Nov 24 09:25:54 crc kubenswrapper[4799]: I1124 09:25:54.805671 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-7h9f4_93716fdc-7850-4f86-af66-6511c476f8b3/ovn-openstack-openstack-cell1/0.log" Nov 24 09:25:54 crc kubenswrapper[4799]: I1124 09:25:54.916793 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2b5580f4-5a36-492a-a9de-cdae05bda4a7/ovsdbserver-nb/0.log" Nov 24 09:25:55 crc kubenswrapper[4799]: I1124 09:25:55.046687 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_f99fe0ba-86b1-47ce-b936-c40ab7e4539e/openstack-network-exporter/0.log" Nov 24 09:25:55 crc kubenswrapper[4799]: I1124 09:25:55.077110 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_f99fe0ba-86b1-47ce-b936-c40ab7e4539e/ovsdbserver-nb/0.log" Nov 24 09:25:55 crc kubenswrapper[4799]: I1124 09:25:55.282741 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_758e3db3-eb2f-48bf-8434-39d0bf4b8595/openstack-network-exporter/0.log" Nov 24 09:25:55 crc kubenswrapper[4799]: I1124 09:25:55.296553 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_758e3db3-eb2f-48bf-8434-39d0bf4b8595/ovsdbserver-nb/0.log" Nov 24 09:25:55 crc kubenswrapper[4799]: I1124 09:25:55.457983 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0ef83552-6d16-407f-a4ed-bd0a8f9c5927/openstack-network-exporter/0.log" Nov 24 09:25:55 crc kubenswrapper[4799]: I1124 09:25:55.494873 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0ef83552-6d16-407f-a4ed-bd0a8f9c5927/ovsdbserver-sb/0.log" Nov 24 09:25:55 crc kubenswrapper[4799]: I1124 09:25:55.701108 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_5703958e-9f91-479c-a89b-c90494eb07ec/openstack-network-exporter/0.log" Nov 24 09:25:55 crc kubenswrapper[4799]: I1124 09:25:55.716417 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_5703958e-9f91-479c-a89b-c90494eb07ec/ovsdbserver-sb/0.log" Nov 24 09:25:55 crc kubenswrapper[4799]: I1124 09:25:55.898434 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_c136e238-389c-490f-a6f5-afa0f00c55dd/openstack-network-exporter/0.log" Nov 24 09:25:55 crc kubenswrapper[4799]: I1124 09:25:55.937800 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_c136e238-389c-490f-a6f5-afa0f00c55dd/ovsdbserver-sb/0.log" Nov 24 09:25:56 crc kubenswrapper[4799]: I1124 09:25:56.207806 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bb89858b4-prpnr_6c4ede8a-4779-4b8e-87c2-9d54664e3f4f/placement-api/0.log" Nov 24 09:25:56 crc kubenswrapper[4799]: I1124 09:25:56.217410 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cxxqfp_de1b21f0-9205-4688-b4d9-dbfc6045dfb1/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Nov 24 09:25:56 crc kubenswrapper[4799]: I1124 09:25:56.283301 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bb89858b4-prpnr_6c4ede8a-4779-4b8e-87c2-9d54664e3f4f/placement-log/0.log" Nov 24 09:25:56 crc kubenswrapper[4799]: I1124 09:25:56.420390 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_d2f7efbf-fb81-4cb1-8284-42baf4c8da5e/init-config-reloader/0.log" Nov 24 09:25:56 crc kubenswrapper[4799]: I1124 09:25:56.692741 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_d2f7efbf-fb81-4cb1-8284-42baf4c8da5e/prometheus/0.log" Nov 24 09:25:56 crc kubenswrapper[4799]: I1124 09:25:56.694061 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_d2f7efbf-fb81-4cb1-8284-42baf4c8da5e/config-reloader/0.log" Nov 24 09:25:56 crc kubenswrapper[4799]: I1124 09:25:56.746033 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_d2f7efbf-fb81-4cb1-8284-42baf4c8da5e/init-config-reloader/0.log" Nov 24 09:25:56 crc kubenswrapper[4799]: I1124 09:25:56.751549 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_d2f7efbf-fb81-4cb1-8284-42baf4c8da5e/thanos-sidecar/0.log" Nov 24 09:25:56 crc kubenswrapper[4799]: I1124 09:25:56.928563 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4/setup-container/0.log" Nov 24 09:25:57 crc kubenswrapper[4799]: I1124 09:25:57.145575 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4/setup-container/0.log" Nov 24 09:25:57 crc kubenswrapper[4799]: I1124 09:25:57.190649 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_2a41cd9d-3e31-4ff5-be57-1cc067c9c0f4/rabbitmq/0.log" Nov 24 09:25:57 crc kubenswrapper[4799]: I1124 09:25:57.200061 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f2bde1fe-98c3-4302-a350-72cfbc703dd2/setup-container/0.log" Nov 24 09:25:57 crc kubenswrapper[4799]: I1124 09:25:57.444768 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f2bde1fe-98c3-4302-a350-72cfbc703dd2/setup-container/0.log" Nov 24 09:25:57 crc kubenswrapper[4799]: I1124 09:25:57.515711 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-mpzh2_92a36210-a0e5-4548-99c1-0bf642f7aeb7/reboot-os-openstack-openstack-cell1/0.log" Nov 24 09:25:57 crc kubenswrapper[4799]: I1124 09:25:57.537804 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f2bde1fe-98c3-4302-a350-72cfbc703dd2/rabbitmq/0.log" Nov 24 09:25:57 crc kubenswrapper[4799]: I1124 09:25:57.763034 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-l6s6d_95317e48-8886-47e4-800e-a6b7e2663eb0/run-os-openstack-openstack-cell1/0.log" Nov 24 09:25:57 crc kubenswrapper[4799]: I1124 09:25:57.817101 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-pr2f5_982af6fd-6fcf-42b7-ae60-1546609a2f8d/ssh-known-hosts-openstack/0.log" Nov 24 09:25:57 crc kubenswrapper[4799]: I1124 09:25:57.994803 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-hvjgw_9459dbfb-9863-41b4-85e3-e18eff5eec82/telemetry-openstack-openstack-cell1/0.log" Nov 24 09:25:58 crc kubenswrapper[4799]: I1124 09:25:58.164081 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_ed0b638f-ace1-45e6-85d4-20ca5ae0ece8/tempest-tests-tempest-tests-runner/0.log" Nov 24 09:25:58 crc kubenswrapper[4799]: I1124 09:25:58.816529 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_dc94072d-adf7-497a-bec7-67b42c411edb/test-operator-logs-container/0.log" Nov 24 09:25:59 crc kubenswrapper[4799]: I1124 09:25:59.197717 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-bz49z_23397601-e699-4796-804a-bcf6b64f4a20/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Nov 24 09:25:59 crc kubenswrapper[4799]: I1124 09:25:59.303382 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-9hl97_525cb867-093d-4379-b617-0590c59f6ef1/validate-network-openstack-openstack-cell1/0.log" Nov 24 09:26:19 crc kubenswrapper[4799]: I1124 09:26:19.397222 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e99ee713-7523-4dde-9751-34c9ef7bad1f/memcached/0.log" Nov 24 09:26:24 crc kubenswrapper[4799]: I1124 09:26:24.949340 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs_3a457c47-44d1-4246-a71b-3c1dc628564e/util/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.153664 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs_3a457c47-44d1-4246-a71b-3c1dc628564e/pull/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.162219 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs_3a457c47-44d1-4246-a71b-3c1dc628564e/util/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.183271 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs_3a457c47-44d1-4246-a71b-3c1dc628564e/pull/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.382823 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs_3a457c47-44d1-4246-a71b-3c1dc628564e/extract/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.388080 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs_3a457c47-44d1-4246-a71b-3c1dc628564e/util/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.405508 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b72876f8bs_3a457c47-44d1-4246-a71b-3c1dc628564e/pull/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.564554 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-kzmqt_f1284cfd-b954-4b2d-8e3c-2e57bb52ca27/kube-rbac-proxy/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.651638 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-w7jld_52251a10-6203-43da-af96-94f78f0edacc/kube-rbac-proxy/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.710509 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-kzmqt_f1284cfd-b954-4b2d-8e3c-2e57bb52ca27/manager/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.855574 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-w7jld_52251a10-6203-43da-af96-94f78f0edacc/manager/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.888489 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-rxlcx_3fec8392-f37a-4f4a-bdec-696bf7b02cd1/kube-rbac-proxy/0.log" Nov 24 09:26:25 crc kubenswrapper[4799]: I1124 09:26:25.973414 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-rxlcx_3fec8392-f37a-4f4a-bdec-696bf7b02cd1/manager/0.log" Nov 24 09:26:26 crc kubenswrapper[4799]: I1124 09:26:26.079826 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-9prrl_4866479c-d0b9-4366-8ab4-f041573d0d35/kube-rbac-proxy/0.log" Nov 24 09:26:26 crc kubenswrapper[4799]: I1124 09:26:26.266306 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-9prrl_4866479c-d0b9-4366-8ab4-f041573d0d35/manager/0.log" Nov 24 09:26:26 crc kubenswrapper[4799]: I1124 09:26:26.400912 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-2zmvd_af65b87f-5fe7-4ade-873b-e4d7040df219/kube-rbac-proxy/0.log" Nov 24 09:26:26 crc kubenswrapper[4799]: I1124 09:26:26.430962 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-2zmvd_af65b87f-5fe7-4ade-873b-e4d7040df219/manager/0.log" Nov 24 09:26:26 crc kubenswrapper[4799]: I1124 09:26:26.506299 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-9x9d2_9f756dcc-daa5-4ed3-a4af-3bbf3c9f186c/kube-rbac-proxy/0.log" Nov 24 09:26:26 crc kubenswrapper[4799]: I1124 09:26:26.626720 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-9x9d2_9f756dcc-daa5-4ed3-a4af-3bbf3c9f186c/manager/0.log" Nov 24 09:26:26 crc kubenswrapper[4799]: I1124 09:26:26.734411 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-s7z8n_9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c/kube-rbac-proxy/0.log" Nov 24 09:26:26 crc kubenswrapper[4799]: I1124 09:26:26.871181 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-mjdkn_d6a55aa5-3e1c-49e9-bd56-48f8497c1b49/kube-rbac-proxy/0.log" Nov 24 09:26:26 crc kubenswrapper[4799]: I1124 09:26:26.956723 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-s7z8n_9f5bb2cc-c16d-42fb-b88e-8bbac7d8b64c/manager/0.log" Nov 24 09:26:26 crc kubenswrapper[4799]: I1124 09:26:26.987748 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-mjdkn_d6a55aa5-3e1c-49e9-bd56-48f8497c1b49/manager/0.log" Nov 24 09:26:27 crc kubenswrapper[4799]: I1124 09:26:27.153618 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-nwftw_32b50dc4-f35a-419b-9a18-9ff05f4c826c/kube-rbac-proxy/0.log" Nov 24 09:26:27 crc kubenswrapper[4799]: I1124 09:26:27.279262 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-nwftw_32b50dc4-f35a-419b-9a18-9ff05f4c826c/manager/0.log" Nov 24 09:26:27 crc kubenswrapper[4799]: I1124 09:26:27.319463 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-2lccf_95afe811-a25f-4b89-9928-ee261693091b/kube-rbac-proxy/0.log" Nov 24 09:26:27 crc kubenswrapper[4799]: I1124 09:26:27.387543 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-2lccf_95afe811-a25f-4b89-9928-ee261693091b/manager/0.log" Nov 24 09:26:27 crc kubenswrapper[4799]: I1124 09:26:27.487955 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-kxlc9_6e6f9937-437a-4f97-a8ea-e486ec9a448c/kube-rbac-proxy/0.log" Nov 24 09:26:27 crc kubenswrapper[4799]: I1124 09:26:27.562209 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-kxlc9_6e6f9937-437a-4f97-a8ea-e486ec9a448c/manager/0.log" Nov 24 09:26:27 crc kubenswrapper[4799]: I1124 09:26:27.698677 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-xgq2g_8a1443d1-4bea-481d-ad92-c92e7f5ac7cd/kube-rbac-proxy/0.log" Nov 24 09:26:27 crc kubenswrapper[4799]: I1124 09:26:27.834467 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-5v5lr_3b5de07d-ee5e-492a-9b9d-26de9837ca80/kube-rbac-proxy/0.log" Nov 24 09:26:27 crc kubenswrapper[4799]: I1124 09:26:27.862029 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-xgq2g_8a1443d1-4bea-481d-ad92-c92e7f5ac7cd/manager/0.log" Nov 24 09:26:28 crc kubenswrapper[4799]: I1124 09:26:28.066164 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-k7jkc_0fce00df-734a-4701-9f85-034b62ec61a9/kube-rbac-proxy/0.log" Nov 24 09:26:28 crc kubenswrapper[4799]: I1124 09:26:28.099764 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-k7jkc_0fce00df-734a-4701-9f85-034b62ec61a9/manager/0.log" Nov 24 09:26:28 crc kubenswrapper[4799]: I1124 09:26:28.142670 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-5v5lr_3b5de07d-ee5e-492a-9b9d-26de9837ca80/manager/0.log" Nov 24 09:26:28 crc kubenswrapper[4799]: I1124 09:26:28.294236 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5_b0be97de-cd56-47c0-bd50-868a8fba50d6/kube-rbac-proxy/0.log" Nov 24 09:26:28 crc kubenswrapper[4799]: I1124 09:26:28.302321 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44lxpb5_b0be97de-cd56-47c0-bd50-868a8fba50d6/manager/0.log" Nov 24 09:26:28 crc kubenswrapper[4799]: I1124 09:26:28.462363 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-px9nk_5c91af0a-fe03-45c5-ae6f-3ea056c17e31/kube-rbac-proxy/0.log" Nov 24 09:26:28 crc kubenswrapper[4799]: I1124 09:26:28.558007 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-g6rfm_98ea352f-7485-4cdb-9ec9-bb208357fefe/kube-rbac-proxy/0.log" Nov 24 09:26:28 crc kubenswrapper[4799]: I1124 09:26:28.852775 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-g6rfm_98ea352f-7485-4cdb-9ec9-bb208357fefe/operator/0.log" Nov 24 09:26:29 crc kubenswrapper[4799]: I1124 09:26:29.125516 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-4k5qm_3d678199-c527-4db6-8e3a-39ba2e59c03a/registry-server/0.log" Nov 24 09:26:29 crc kubenswrapper[4799]: I1124 09:26:29.148754 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-2xzbc_fc6a7978-e9a3-4263-89d9-7ed03da8035f/kube-rbac-proxy/0.log" Nov 24 09:26:29 crc kubenswrapper[4799]: I1124 09:26:29.395693 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-74qt8_9d9e84d5-49ab-46aa-971e-cd920e2801b1/manager/0.log" Nov 24 09:26:29 crc kubenswrapper[4799]: I1124 09:26:29.400167 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-74qt8_9d9e84d5-49ab-46aa-971e-cd920e2801b1/kube-rbac-proxy/0.log" Nov 24 09:26:29 crc kubenswrapper[4799]: I1124 09:26:29.404016 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-2xzbc_fc6a7978-e9a3-4263-89d9-7ed03da8035f/manager/0.log" Nov 24 09:26:29 crc kubenswrapper[4799]: I1124 09:26:29.690744 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-gv46d_611f4120-8c1b-4745-a5d7-4e1dd52ec567/operator/0.log" Nov 24 09:26:29 crc kubenswrapper[4799]: I1124 09:26:29.692141 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-scbrc_66c8f0b0-71e8-4bf6-a28b-109bff76c35e/kube-rbac-proxy/0.log" Nov 24 09:26:29 crc kubenswrapper[4799]: I1124 09:26:29.862894 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-scbrc_66c8f0b0-71e8-4bf6-a28b-109bff76c35e/manager/0.log" Nov 24 09:26:29 crc kubenswrapper[4799]: I1124 09:26:29.878612 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-d4b8t_1d093c2b-22f5-4146-a0e1-a29ab68e1ab3/kube-rbac-proxy/0.log" Nov 24 09:26:30 crc kubenswrapper[4799]: I1124 09:26:30.102836 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-d4b8t_1d093c2b-22f5-4146-a0e1-a29ab68e1ab3/manager/0.log" Nov 24 09:26:30 crc kubenswrapper[4799]: I1124 09:26:30.174997 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-9rph7_486bc897-b77f-41d9-8507-a667d595eb61/kube-rbac-proxy/0.log" Nov 24 09:26:30 crc kubenswrapper[4799]: I1124 09:26:30.284623 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-9rph7_486bc897-b77f-41d9-8507-a667d595eb61/manager/0.log" Nov 24 09:26:30 crc kubenswrapper[4799]: I1124 09:26:30.366861 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-fghbt_9ecccf0c-10f4-43e6-b658-528ff9acea0f/kube-rbac-proxy/0.log" Nov 24 09:26:30 crc kubenswrapper[4799]: I1124 09:26:30.407214 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-fghbt_9ecccf0c-10f4-43e6-b658-528ff9acea0f/manager/0.log" Nov 24 09:26:30 crc kubenswrapper[4799]: I1124 09:26:30.861927 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-px9nk_5c91af0a-fe03-45c5-ae6f-3ea056c17e31/manager/0.log" Nov 24 09:26:48 crc kubenswrapper[4799]: I1124 09:26:48.930714 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-knplv_fde84f5d-c045-4a70-9d7d-9f4ca71d4351/control-plane-machine-set-operator/0.log" Nov 24 09:26:49 crc kubenswrapper[4799]: I1124 09:26:49.169020 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-t6hw2_86f640cf-7c54-4b44-a52f-21378c535d5a/kube-rbac-proxy/0.log" Nov 24 09:26:49 crc kubenswrapper[4799]: I1124 09:26:49.179294 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-t6hw2_86f640cf-7c54-4b44-a52f-21378c535d5a/machine-api-operator/0.log" Nov 24 09:26:50 crc kubenswrapper[4799]: I1124 09:26:50.401214 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:26:50 crc kubenswrapper[4799]: I1124 09:26:50.401798 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:27:02 crc kubenswrapper[4799]: I1124 09:27:02.626161 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-btwn9_71cb8206-339b-4aff-aa85-d009840a5fd6/cert-manager-controller/0.log" Nov 24 09:27:02 crc kubenswrapper[4799]: I1124 09:27:02.830271 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-ktx6z_439904d9-d1e0-4733-8f08-db0f792fdcc2/cert-manager-cainjector/0.log" Nov 24 09:27:03 crc kubenswrapper[4799]: I1124 09:27:03.577566 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-58pnr_a45abe66-dc32-4b76-90d9-b821dba28bd5/cert-manager-webhook/0.log" Nov 24 09:27:16 crc kubenswrapper[4799]: I1124 09:27:16.539897 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-kz4r6_0a5a4df1-9ab8-4b31-b31d-077678cb993f/nmstate-console-plugin/0.log" Nov 24 09:27:16 crc kubenswrapper[4799]: I1124 09:27:16.722688 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-p9xqt_4f03b272-18b2-423e-9e5d-da5fb982df9c/nmstate-handler/0.log" Nov 24 09:27:16 crc kubenswrapper[4799]: I1124 09:27:16.728759 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-56p9j_d960de55-1758-407b-a51d-62fa10fc28dd/kube-rbac-proxy/0.log" Nov 24 09:27:16 crc kubenswrapper[4799]: I1124 09:27:16.797641 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-56p9j_d960de55-1758-407b-a51d-62fa10fc28dd/nmstate-metrics/0.log" Nov 24 09:27:16 crc kubenswrapper[4799]: I1124 09:27:16.985315 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-8swlg_d89cee1d-8190-4238-a6ae-f8b629e40cc0/nmstate-operator/0.log" Nov 24 09:27:17 crc kubenswrapper[4799]: I1124 09:27:17.041262 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-h7t27_72c88d6c-d53a-47e1-a1e2-4b7292c814b4/nmstate-webhook/0.log" Nov 24 09:27:20 crc kubenswrapper[4799]: I1124 09:27:20.400219 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:27:20 crc kubenswrapper[4799]: I1124 09:27:20.400508 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:27:31 crc kubenswrapper[4799]: I1124 09:27:31.294951 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-gkkl2_14418811-1eac-4ac6-a598-8bde1dc50a82/kube-rbac-proxy/0.log" Nov 24 09:27:31 crc kubenswrapper[4799]: I1124 09:27:31.620306 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-frr-files/0.log" Nov 24 09:27:31 crc kubenswrapper[4799]: I1124 09:27:31.803525 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-gkkl2_14418811-1eac-4ac6-a598-8bde1dc50a82/controller/0.log" Nov 24 09:27:31 crc kubenswrapper[4799]: I1124 09:27:31.912521 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-metrics/0.log" Nov 24 09:27:31 crc kubenswrapper[4799]: I1124 09:27:31.931969 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-frr-files/0.log" Nov 24 09:27:31 crc kubenswrapper[4799]: I1124 09:27:31.992626 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-reloader/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.031813 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-reloader/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.190212 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-reloader/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.212653 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-metrics/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.215305 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-frr-files/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.245411 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-metrics/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.424116 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-frr-files/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.438382 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-reloader/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.463458 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/cp-metrics/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.509295 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/controller/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.623687 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/frr-metrics/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.699709 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/kube-rbac-proxy/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.784747 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/kube-rbac-proxy-frr/0.log" Nov 24 09:27:32 crc kubenswrapper[4799]: I1124 09:27:32.852769 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/reloader/0.log" Nov 24 09:27:33 crc kubenswrapper[4799]: I1124 09:27:33.021814 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-7khps_6cd58852-5a3f-46aa-b43a-6a8ab49711a5/frr-k8s-webhook-server/0.log" Nov 24 09:27:33 crc kubenswrapper[4799]: I1124 09:27:33.211729 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-66f4d4788d-6qmmx_ead0c22c-848d-4391-903b-62aa4bbc697c/manager/0.log" Nov 24 09:27:33 crc kubenswrapper[4799]: I1124 09:27:33.385988 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-666db98b56-rbfxp_5fcab455-bba7-4f3c-84c7-8e938d1ce6cd/webhook-server/0.log" Nov 24 09:27:33 crc kubenswrapper[4799]: I1124 09:27:33.523037 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bzp2s_8a7ba0ee-de14-43e5-8b93-9b915e8d9938/kube-rbac-proxy/0.log" Nov 24 09:27:34 crc kubenswrapper[4799]: I1124 09:27:34.651641 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bzp2s_8a7ba0ee-de14-43e5-8b93-9b915e8d9938/speaker/0.log" Nov 24 09:27:36 crc kubenswrapper[4799]: I1124 09:27:36.469526 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2gqq5_aa96ad1b-f53e-428e-ac55-bb15c9740c26/frr/0.log" Nov 24 09:27:47 crc kubenswrapper[4799]: I1124 09:27:47.571528 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb_df363f0a-38fa-4510-9a1a-947e628b5f89/util/0.log" Nov 24 09:27:47 crc kubenswrapper[4799]: I1124 09:27:47.728991 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb_df363f0a-38fa-4510-9a1a-947e628b5f89/util/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.191817 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb_df363f0a-38fa-4510-9a1a-947e628b5f89/pull/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.193016 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb_df363f0a-38fa-4510-9a1a-947e628b5f89/pull/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.248261 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb_df363f0a-38fa-4510-9a1a-947e628b5f89/pull/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.411502 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb_df363f0a-38fa-4510-9a1a-947e628b5f89/extract/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.425210 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a5nxfb_df363f0a-38fa-4510-9a1a-947e628b5f89/util/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.458995 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr_e055d22a-1e71-499f-a898-fdcff0b33a35/util/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.687362 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr_e055d22a-1e71-499f-a898-fdcff0b33a35/util/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.715860 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr_e055d22a-1e71-499f-a898-fdcff0b33a35/pull/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.723238 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr_e055d22a-1e71-499f-a898-fdcff0b33a35/pull/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.887374 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr_e055d22a-1e71-499f-a898-fdcff0b33a35/util/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.896952 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr_e055d22a-1e71-499f-a898-fdcff0b33a35/pull/0.log" Nov 24 09:27:48 crc kubenswrapper[4799]: I1124 09:27:48.901631 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772elz4tr_e055d22a-1e71-499f-a898-fdcff0b33a35/extract/0.log" Nov 24 09:27:49 crc kubenswrapper[4799]: I1124 09:27:49.067603 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x_b051e98e-b756-4df6-ba65-29e4ac6572ca/util/0.log" Nov 24 09:27:49 crc kubenswrapper[4799]: I1124 09:27:49.203312 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x_b051e98e-b756-4df6-ba65-29e4ac6572ca/pull/0.log" Nov 24 09:27:49 crc kubenswrapper[4799]: I1124 09:27:49.233243 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x_b051e98e-b756-4df6-ba65-29e4ac6572ca/util/0.log" Nov 24 09:27:49 crc kubenswrapper[4799]: I1124 09:27:49.283827 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x_b051e98e-b756-4df6-ba65-29e4ac6572ca/pull/0.log" Nov 24 09:27:49 crc kubenswrapper[4799]: I1124 09:27:49.433866 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x_b051e98e-b756-4df6-ba65-29e4ac6572ca/util/0.log" Nov 24 09:27:49 crc kubenswrapper[4799]: I1124 09:27:49.437995 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x_b051e98e-b756-4df6-ba65-29e4ac6572ca/pull/0.log" Nov 24 09:27:49 crc kubenswrapper[4799]: I1124 09:27:49.456987 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210cmv6x_b051e98e-b756-4df6-ba65-29e4ac6572ca/extract/0.log" Nov 24 09:27:49 crc kubenswrapper[4799]: I1124 09:27:49.633695 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s2vvw_b7277667-0bca-4cb2-9e44-304597d405c8/extract-utilities/0.log" Nov 24 09:27:49 crc kubenswrapper[4799]: I1124 09:27:49.849772 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s2vvw_b7277667-0bca-4cb2-9e44-304597d405c8/extract-utilities/0.log" Nov 24 09:27:49 crc kubenswrapper[4799]: I1124 09:27:49.876173 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s2vvw_b7277667-0bca-4cb2-9e44-304597d405c8/extract-content/0.log" Nov 24 09:27:49 crc kubenswrapper[4799]: I1124 09:27:49.896662 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s2vvw_b7277667-0bca-4cb2-9e44-304597d405c8/extract-content/0.log" Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.064583 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s2vvw_b7277667-0bca-4cb2-9e44-304597d405c8/extract-utilities/0.log" Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.064672 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s2vvw_b7277667-0bca-4cb2-9e44-304597d405c8/extract-content/0.log" Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.316902 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rcrjl_7ee74762-ac3f-472e-9054-529e6b3a293f/extract-utilities/0.log" Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.400366 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.400645 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.400755 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.401547 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"517c10c0986c1088f2ae888718c16d2c685822643c593b40da162bf2f3673af3"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.401670 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://517c10c0986c1088f2ae888718c16d2c685822643c593b40da162bf2f3673af3" gracePeriod=600 Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.529915 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rcrjl_7ee74762-ac3f-472e-9054-529e6b3a293f/extract-utilities/0.log" Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.579577 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rcrjl_7ee74762-ac3f-472e-9054-529e6b3a293f/extract-content/0.log" Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.597646 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rcrjl_7ee74762-ac3f-472e-9054-529e6b3a293f/extract-content/0.log" Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.752288 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rcrjl_7ee74762-ac3f-472e-9054-529e6b3a293f/extract-utilities/0.log" Nov 24 09:27:50 crc kubenswrapper[4799]: I1124 09:27:50.930806 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rcrjl_7ee74762-ac3f-472e-9054-529e6b3a293f/extract-content/0.log" Nov 24 09:27:51 crc kubenswrapper[4799]: I1124 09:27:51.136337 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x_7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0/util/0.log" Nov 24 09:27:51 crc kubenswrapper[4799]: I1124 09:27:51.239445 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="517c10c0986c1088f2ae888718c16d2c685822643c593b40da162bf2f3673af3" exitCode=0 Nov 24 09:27:51 crc kubenswrapper[4799]: I1124 09:27:51.239504 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"517c10c0986c1088f2ae888718c16d2c685822643c593b40da162bf2f3673af3"} Nov 24 09:27:51 crc kubenswrapper[4799]: I1124 09:27:51.239538 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerStarted","Data":"626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532"} Nov 24 09:27:51 crc kubenswrapper[4799]: I1124 09:27:51.239562 4799 scope.go:117] "RemoveContainer" containerID="0828281e498f16ac323844cb1af3056fc20ffc87b59ee18875e98fa78d9dfee2" Nov 24 09:27:51 crc kubenswrapper[4799]: I1124 09:27:51.640452 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s2vvw_b7277667-0bca-4cb2-9e44-304597d405c8/registry-server/0.log" Nov 24 09:27:51 crc kubenswrapper[4799]: I1124 09:27:51.670498 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x_7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0/pull/0.log" Nov 24 09:27:51 crc kubenswrapper[4799]: I1124 09:27:51.730330 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x_7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0/util/0.log" Nov 24 09:27:51 crc kubenswrapper[4799]: I1124 09:27:51.733794 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x_7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0/pull/0.log" Nov 24 09:27:51 crc kubenswrapper[4799]: I1124 09:27:51.973469 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x_7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0/util/0.log" Nov 24 09:27:52 crc kubenswrapper[4799]: I1124 09:27:52.012670 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x_7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0/extract/0.log" Nov 24 09:27:52 crc kubenswrapper[4799]: I1124 09:27:52.087748 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mq26x_7a5708a5-e3c4-4b0b-9dbb-1e4bf8ed7da0/pull/0.log" Nov 24 09:27:52 crc kubenswrapper[4799]: I1124 09:27:52.219501 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2hdxh_3c598ccb-346a-45eb-ad78-ad538667a705/marketplace-operator/0.log" Nov 24 09:27:52 crc kubenswrapper[4799]: I1124 09:27:52.302357 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r5bgj_15f64dad-142a-42ca-b7c0-c4adeaec75ee/extract-utilities/0.log" Nov 24 09:27:52 crc kubenswrapper[4799]: I1124 09:27:52.397401 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rcrjl_7ee74762-ac3f-472e-9054-529e6b3a293f/registry-server/0.log" Nov 24 09:27:52 crc kubenswrapper[4799]: I1124 09:27:52.531755 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r5bgj_15f64dad-142a-42ca-b7c0-c4adeaec75ee/extract-utilities/0.log" Nov 24 09:27:52 crc kubenswrapper[4799]: I1124 09:27:52.544794 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r5bgj_15f64dad-142a-42ca-b7c0-c4adeaec75ee/extract-content/0.log" Nov 24 09:27:52 crc kubenswrapper[4799]: I1124 09:27:52.565148 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r5bgj_15f64dad-142a-42ca-b7c0-c4adeaec75ee/extract-content/0.log" Nov 24 09:27:52 crc kubenswrapper[4799]: I1124 09:27:52.766725 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r5bgj_15f64dad-142a-42ca-b7c0-c4adeaec75ee/extract-content/0.log" Nov 24 09:27:52 crc kubenswrapper[4799]: I1124 09:27:52.776243 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r5bgj_15f64dad-142a-42ca-b7c0-c4adeaec75ee/extract-utilities/0.log" Nov 24 09:27:52 crc kubenswrapper[4799]: I1124 09:27:52.800558 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c5dq6_40eb2d35-dc7b-4e12-860a-e13da529af4c/extract-utilities/0.log" Nov 24 09:27:53 crc kubenswrapper[4799]: I1124 09:27:53.046498 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c5dq6_40eb2d35-dc7b-4e12-860a-e13da529af4c/extract-content/0.log" Nov 24 09:27:53 crc kubenswrapper[4799]: I1124 09:27:53.091171 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c5dq6_40eb2d35-dc7b-4e12-860a-e13da529af4c/extract-content/0.log" Nov 24 09:27:53 crc kubenswrapper[4799]: I1124 09:27:53.122399 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r5bgj_15f64dad-142a-42ca-b7c0-c4adeaec75ee/registry-server/0.log" Nov 24 09:27:53 crc kubenswrapper[4799]: I1124 09:27:53.123300 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c5dq6_40eb2d35-dc7b-4e12-860a-e13da529af4c/extract-utilities/0.log" Nov 24 09:27:53 crc kubenswrapper[4799]: I1124 09:27:53.274984 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c5dq6_40eb2d35-dc7b-4e12-860a-e13da529af4c/extract-utilities/0.log" Nov 24 09:27:53 crc kubenswrapper[4799]: I1124 09:27:53.286473 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c5dq6_40eb2d35-dc7b-4e12-860a-e13da529af4c/extract-content/0.log" Nov 24 09:27:53 crc kubenswrapper[4799]: I1124 09:27:53.413804 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c5dq6_40eb2d35-dc7b-4e12-860a-e13da529af4c/registry-server/0.log" Nov 24 09:28:09 crc kubenswrapper[4799]: I1124 09:28:09.060674 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-ftn2m_f6f9c83e-09c2-43c5-bb06-a0ea47981131/prometheus-operator/0.log" Nov 24 09:28:09 crc kubenswrapper[4799]: I1124 09:28:09.273382 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-86db4855d-j9twl_0b438aea-0c2b-41eb-8e2d-8b77fa4a7574/prometheus-operator-admission-webhook/0.log" Nov 24 09:28:09 crc kubenswrapper[4799]: I1124 09:28:09.299630 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-86db4855d-vb9r2_5ac7f1a5-756a-4dab-bde7-a68e99f05ed1/prometheus-operator-admission-webhook/0.log" Nov 24 09:28:09 crc kubenswrapper[4799]: I1124 09:28:09.494664 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-wnwkw_3a9f5a89-4ea2-4106-86fe-8bcc9b7f6820/operator/0.log" Nov 24 09:28:09 crc kubenswrapper[4799]: I1124 09:28:09.598609 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-26xdx_26315ae1-10ac-4122-b2ac-6ae36c41fd84/perses-operator/0.log" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.383838 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wq6ks"] Nov 24 09:28:33 crc kubenswrapper[4799]: E1124 09:28:33.384868 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee4b3a4-a03f-425a-9220-fa3c378400b0" containerName="extract-content" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.384882 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee4b3a4-a03f-425a-9220-fa3c378400b0" containerName="extract-content" Nov 24 09:28:33 crc kubenswrapper[4799]: E1124 09:28:33.384897 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee4b3a4-a03f-425a-9220-fa3c378400b0" containerName="registry-server" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.384904 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee4b3a4-a03f-425a-9220-fa3c378400b0" containerName="registry-server" Nov 24 09:28:33 crc kubenswrapper[4799]: E1124 09:28:33.384939 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee4b3a4-a03f-425a-9220-fa3c378400b0" containerName="extract-utilities" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.384946 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee4b3a4-a03f-425a-9220-fa3c378400b0" containerName="extract-utilities" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.385203 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="eee4b3a4-a03f-425a-9220-fa3c378400b0" containerName="registry-server" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.386821 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.398990 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wq6ks"] Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.532057 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-utilities\") pod \"redhat-marketplace-wq6ks\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.532193 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-catalog-content\") pod \"redhat-marketplace-wq6ks\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.532295 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b75dk\" (UniqueName: \"kubernetes.io/projected/f19a5363-2345-4467-a77f-412cda120b3f-kube-api-access-b75dk\") pod \"redhat-marketplace-wq6ks\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.633900 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-utilities\") pod \"redhat-marketplace-wq6ks\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.634000 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-catalog-content\") pod \"redhat-marketplace-wq6ks\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.634056 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b75dk\" (UniqueName: \"kubernetes.io/projected/f19a5363-2345-4467-a77f-412cda120b3f-kube-api-access-b75dk\") pod \"redhat-marketplace-wq6ks\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.635313 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-utilities\") pod \"redhat-marketplace-wq6ks\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.635363 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-catalog-content\") pod \"redhat-marketplace-wq6ks\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.656119 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b75dk\" (UniqueName: \"kubernetes.io/projected/f19a5363-2345-4467-a77f-412cda120b3f-kube-api-access-b75dk\") pod \"redhat-marketplace-wq6ks\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:33 crc kubenswrapper[4799]: I1124 09:28:33.735567 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:34 crc kubenswrapper[4799]: I1124 09:28:34.275135 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wq6ks"] Nov 24 09:28:34 crc kubenswrapper[4799]: W1124 09:28:34.289841 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf19a5363_2345_4467_a77f_412cda120b3f.slice/crio-bfabcd3ad9c95062388b3fca7ef860ed72a5b384d753ad6067f5176240997d39 WatchSource:0}: Error finding container bfabcd3ad9c95062388b3fca7ef860ed72a5b384d753ad6067f5176240997d39: Status 404 returned error can't find the container with id bfabcd3ad9c95062388b3fca7ef860ed72a5b384d753ad6067f5176240997d39 Nov 24 09:28:34 crc kubenswrapper[4799]: I1124 09:28:34.750896 4799 generic.go:334] "Generic (PLEG): container finished" podID="f19a5363-2345-4467-a77f-412cda120b3f" containerID="f232ffb7490dffcc30ba61567c7e9b104426b05a858c77699b31bca9d9a79dbb" exitCode=0 Nov 24 09:28:34 crc kubenswrapper[4799]: I1124 09:28:34.750947 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wq6ks" event={"ID":"f19a5363-2345-4467-a77f-412cda120b3f","Type":"ContainerDied","Data":"f232ffb7490dffcc30ba61567c7e9b104426b05a858c77699b31bca9d9a79dbb"} Nov 24 09:28:34 crc kubenswrapper[4799]: I1124 09:28:34.750982 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wq6ks" event={"ID":"f19a5363-2345-4467-a77f-412cda120b3f","Type":"ContainerStarted","Data":"bfabcd3ad9c95062388b3fca7ef860ed72a5b384d753ad6067f5176240997d39"} Nov 24 09:28:35 crc kubenswrapper[4799]: I1124 09:28:35.764364 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wq6ks" event={"ID":"f19a5363-2345-4467-a77f-412cda120b3f","Type":"ContainerStarted","Data":"2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0"} Nov 24 09:28:36 crc kubenswrapper[4799]: I1124 09:28:36.776561 4799 generic.go:334] "Generic (PLEG): container finished" podID="f19a5363-2345-4467-a77f-412cda120b3f" containerID="2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0" exitCode=0 Nov 24 09:28:36 crc kubenswrapper[4799]: I1124 09:28:36.777040 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wq6ks" event={"ID":"f19a5363-2345-4467-a77f-412cda120b3f","Type":"ContainerDied","Data":"2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0"} Nov 24 09:28:37 crc kubenswrapper[4799]: I1124 09:28:37.787674 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wq6ks" event={"ID":"f19a5363-2345-4467-a77f-412cda120b3f","Type":"ContainerStarted","Data":"cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c"} Nov 24 09:28:37 crc kubenswrapper[4799]: I1124 09:28:37.808893 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wq6ks" podStartSLOduration=2.319677886 podStartE2EDuration="4.808875407s" podCreationTimestamp="2025-11-24 09:28:33 +0000 UTC" firstStartedPulling="2025-11-24 09:28:34.753462102 +0000 UTC m=+9660.409444586" lastFinishedPulling="2025-11-24 09:28:37.242659623 +0000 UTC m=+9662.898642107" observedRunningTime="2025-11-24 09:28:37.805562132 +0000 UTC m=+9663.461544636" watchObservedRunningTime="2025-11-24 09:28:37.808875407 +0000 UTC m=+9663.464857881" Nov 24 09:28:43 crc kubenswrapper[4799]: I1124 09:28:43.736257 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:43 crc kubenswrapper[4799]: I1124 09:28:43.736888 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:43 crc kubenswrapper[4799]: I1124 09:28:43.791494 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:43 crc kubenswrapper[4799]: I1124 09:28:43.899765 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:44 crc kubenswrapper[4799]: I1124 09:28:44.030357 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wq6ks"] Nov 24 09:28:45 crc kubenswrapper[4799]: I1124 09:28:45.863747 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wq6ks" podUID="f19a5363-2345-4467-a77f-412cda120b3f" containerName="registry-server" containerID="cri-o://cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c" gracePeriod=2 Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.393125 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.503333 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-utilities\") pod \"f19a5363-2345-4467-a77f-412cda120b3f\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.503786 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-catalog-content\") pod \"f19a5363-2345-4467-a77f-412cda120b3f\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.504141 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b75dk\" (UniqueName: \"kubernetes.io/projected/f19a5363-2345-4467-a77f-412cda120b3f-kube-api-access-b75dk\") pod \"f19a5363-2345-4467-a77f-412cda120b3f\" (UID: \"f19a5363-2345-4467-a77f-412cda120b3f\") " Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.504541 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-utilities" (OuterVolumeSpecName: "utilities") pod "f19a5363-2345-4467-a77f-412cda120b3f" (UID: "f19a5363-2345-4467-a77f-412cda120b3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.505030 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.510633 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f19a5363-2345-4467-a77f-412cda120b3f-kube-api-access-b75dk" (OuterVolumeSpecName: "kube-api-access-b75dk") pod "f19a5363-2345-4467-a77f-412cda120b3f" (UID: "f19a5363-2345-4467-a77f-412cda120b3f"). InnerVolumeSpecName "kube-api-access-b75dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.523167 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f19a5363-2345-4467-a77f-412cda120b3f" (UID: "f19a5363-2345-4467-a77f-412cda120b3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.606944 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b75dk\" (UniqueName: \"kubernetes.io/projected/f19a5363-2345-4467-a77f-412cda120b3f-kube-api-access-b75dk\") on node \"crc\" DevicePath \"\"" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.606979 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f19a5363-2345-4467-a77f-412cda120b3f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.875202 4799 generic.go:334] "Generic (PLEG): container finished" podID="f19a5363-2345-4467-a77f-412cda120b3f" containerID="cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c" exitCode=0 Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.875270 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wq6ks" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.875283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wq6ks" event={"ID":"f19a5363-2345-4467-a77f-412cda120b3f","Type":"ContainerDied","Data":"cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c"} Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.875713 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wq6ks" event={"ID":"f19a5363-2345-4467-a77f-412cda120b3f","Type":"ContainerDied","Data":"bfabcd3ad9c95062388b3fca7ef860ed72a5b384d753ad6067f5176240997d39"} Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.875736 4799 scope.go:117] "RemoveContainer" containerID="cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.896968 4799 scope.go:117] "RemoveContainer" containerID="2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.920705 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wq6ks"] Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.931498 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wq6ks"] Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.933015 4799 scope.go:117] "RemoveContainer" containerID="f232ffb7490dffcc30ba61567c7e9b104426b05a858c77699b31bca9d9a79dbb" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.975160 4799 scope.go:117] "RemoveContainer" containerID="cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c" Nov 24 09:28:46 crc kubenswrapper[4799]: E1124 09:28:46.975760 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c\": container with ID starting with cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c not found: ID does not exist" containerID="cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.975818 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c"} err="failed to get container status \"cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c\": rpc error: code = NotFound desc = could not find container \"cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c\": container with ID starting with cb4366083b3311a6fe662f1aaa1455f16b3819a14649337521637414043d5b1c not found: ID does not exist" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.975872 4799 scope.go:117] "RemoveContainer" containerID="2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0" Nov 24 09:28:46 crc kubenswrapper[4799]: E1124 09:28:46.976311 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0\": container with ID starting with 2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0 not found: ID does not exist" containerID="2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.976343 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0"} err="failed to get container status \"2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0\": rpc error: code = NotFound desc = could not find container \"2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0\": container with ID starting with 2e6bc0a147abe67935c0a0967d30ca8e6eb9b4a15b1b46c3227cb58fa4eddab0 not found: ID does not exist" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.976363 4799 scope.go:117] "RemoveContainer" containerID="f232ffb7490dffcc30ba61567c7e9b104426b05a858c77699b31bca9d9a79dbb" Nov 24 09:28:46 crc kubenswrapper[4799]: E1124 09:28:46.977063 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f232ffb7490dffcc30ba61567c7e9b104426b05a858c77699b31bca9d9a79dbb\": container with ID starting with f232ffb7490dffcc30ba61567c7e9b104426b05a858c77699b31bca9d9a79dbb not found: ID does not exist" containerID="f232ffb7490dffcc30ba61567c7e9b104426b05a858c77699b31bca9d9a79dbb" Nov 24 09:28:46 crc kubenswrapper[4799]: I1124 09:28:46.977117 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f232ffb7490dffcc30ba61567c7e9b104426b05a858c77699b31bca9d9a79dbb"} err="failed to get container status \"f232ffb7490dffcc30ba61567c7e9b104426b05a858c77699b31bca9d9a79dbb\": rpc error: code = NotFound desc = could not find container \"f232ffb7490dffcc30ba61567c7e9b104426b05a858c77699b31bca9d9a79dbb\": container with ID starting with f232ffb7490dffcc30ba61567c7e9b104426b05a858c77699b31bca9d9a79dbb not found: ID does not exist" Nov 24 09:28:47 crc kubenswrapper[4799]: I1124 09:28:47.641277 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f19a5363-2345-4467-a77f-412cda120b3f" path="/var/lib/kubelet/pods/f19a5363-2345-4467-a77f-412cda120b3f/volumes" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.546722 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xs5gp"] Nov 24 09:29:00 crc kubenswrapper[4799]: E1124 09:29:00.547762 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f19a5363-2345-4467-a77f-412cda120b3f" containerName="registry-server" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.547781 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f19a5363-2345-4467-a77f-412cda120b3f" containerName="registry-server" Nov 24 09:29:00 crc kubenswrapper[4799]: E1124 09:29:00.547804 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f19a5363-2345-4467-a77f-412cda120b3f" containerName="extract-utilities" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.547812 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f19a5363-2345-4467-a77f-412cda120b3f" containerName="extract-utilities" Nov 24 09:29:00 crc kubenswrapper[4799]: E1124 09:29:00.547825 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f19a5363-2345-4467-a77f-412cda120b3f" containerName="extract-content" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.547833 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f19a5363-2345-4467-a77f-412cda120b3f" containerName="extract-content" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.548164 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f19a5363-2345-4467-a77f-412cda120b3f" containerName="registry-server" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.552730 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.562366 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xs5gp"] Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.724572 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-catalog-content\") pod \"redhat-operators-xs5gp\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.724993 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6vs4\" (UniqueName: \"kubernetes.io/projected/c82c0477-5daa-4cce-a62d-05bb4f64f49b-kube-api-access-x6vs4\") pod \"redhat-operators-xs5gp\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.725373 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-utilities\") pod \"redhat-operators-xs5gp\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.826681 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-catalog-content\") pod \"redhat-operators-xs5gp\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.826754 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6vs4\" (UniqueName: \"kubernetes.io/projected/c82c0477-5daa-4cce-a62d-05bb4f64f49b-kube-api-access-x6vs4\") pod \"redhat-operators-xs5gp\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.826948 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-utilities\") pod \"redhat-operators-xs5gp\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.827276 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-catalog-content\") pod \"redhat-operators-xs5gp\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:00 crc kubenswrapper[4799]: I1124 09:29:00.827310 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-utilities\") pod \"redhat-operators-xs5gp\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:01 crc kubenswrapper[4799]: I1124 09:29:01.188903 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6vs4\" (UniqueName: \"kubernetes.io/projected/c82c0477-5daa-4cce-a62d-05bb4f64f49b-kube-api-access-x6vs4\") pod \"redhat-operators-xs5gp\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:01 crc kubenswrapper[4799]: I1124 09:29:01.487942 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:02 crc kubenswrapper[4799]: I1124 09:29:02.014680 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xs5gp"] Nov 24 09:29:02 crc kubenswrapper[4799]: I1124 09:29:02.032707 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs5gp" event={"ID":"c82c0477-5daa-4cce-a62d-05bb4f64f49b","Type":"ContainerStarted","Data":"af55782a8c0139af72633cb852be76e2825417a25238d017e403752cb3b1e7c7"} Nov 24 09:29:03 crc kubenswrapper[4799]: I1124 09:29:03.045083 4799 generic.go:334] "Generic (PLEG): container finished" podID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerID="7abe6ba20d98418c262adcb19e266f84efe5b63249ff9f7fb9d0ae1b22e005e6" exitCode=0 Nov 24 09:29:03 crc kubenswrapper[4799]: I1124 09:29:03.045291 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs5gp" event={"ID":"c82c0477-5daa-4cce-a62d-05bb4f64f49b","Type":"ContainerDied","Data":"7abe6ba20d98418c262adcb19e266f84efe5b63249ff9f7fb9d0ae1b22e005e6"} Nov 24 09:29:05 crc kubenswrapper[4799]: I1124 09:29:05.083745 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs5gp" event={"ID":"c82c0477-5daa-4cce-a62d-05bb4f64f49b","Type":"ContainerStarted","Data":"9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387"} Nov 24 09:29:07 crc kubenswrapper[4799]: I1124 09:29:07.111003 4799 generic.go:334] "Generic (PLEG): container finished" podID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerID="9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387" exitCode=0 Nov 24 09:29:07 crc kubenswrapper[4799]: I1124 09:29:07.111318 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs5gp" event={"ID":"c82c0477-5daa-4cce-a62d-05bb4f64f49b","Type":"ContainerDied","Data":"9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387"} Nov 24 09:29:08 crc kubenswrapper[4799]: I1124 09:29:08.124415 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs5gp" event={"ID":"c82c0477-5daa-4cce-a62d-05bb4f64f49b","Type":"ContainerStarted","Data":"48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964"} Nov 24 09:29:08 crc kubenswrapper[4799]: I1124 09:29:08.149447 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xs5gp" podStartSLOduration=3.429227819 podStartE2EDuration="8.149428721s" podCreationTimestamp="2025-11-24 09:29:00 +0000 UTC" firstStartedPulling="2025-11-24 09:29:03.057203653 +0000 UTC m=+9688.713186137" lastFinishedPulling="2025-11-24 09:29:07.777404565 +0000 UTC m=+9693.433387039" observedRunningTime="2025-11-24 09:29:08.148273918 +0000 UTC m=+9693.804256412" watchObservedRunningTime="2025-11-24 09:29:08.149428721 +0000 UTC m=+9693.805411195" Nov 24 09:29:11 crc kubenswrapper[4799]: I1124 09:29:11.489460 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:11 crc kubenswrapper[4799]: I1124 09:29:11.490204 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:13 crc kubenswrapper[4799]: I1124 09:29:13.008422 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xs5gp" podUID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerName="registry-server" probeResult="failure" output=< Nov 24 09:29:13 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Nov 24 09:29:13 crc kubenswrapper[4799]: > Nov 24 09:29:21 crc kubenswrapper[4799]: I1124 09:29:21.536845 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:21 crc kubenswrapper[4799]: I1124 09:29:21.584354 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:21 crc kubenswrapper[4799]: I1124 09:29:21.788780 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xs5gp"] Nov 24 09:29:23 crc kubenswrapper[4799]: I1124 09:29:23.286722 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xs5gp" podUID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerName="registry-server" containerID="cri-o://48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964" gracePeriod=2 Nov 24 09:29:23 crc kubenswrapper[4799]: I1124 09:29:23.846038 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:23 crc kubenswrapper[4799]: I1124 09:29:23.928142 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6vs4\" (UniqueName: \"kubernetes.io/projected/c82c0477-5daa-4cce-a62d-05bb4f64f49b-kube-api-access-x6vs4\") pod \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " Nov 24 09:29:23 crc kubenswrapper[4799]: I1124 09:29:23.928274 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-catalog-content\") pod \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " Nov 24 09:29:23 crc kubenswrapper[4799]: I1124 09:29:23.928335 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-utilities\") pod \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\" (UID: \"c82c0477-5daa-4cce-a62d-05bb4f64f49b\") " Nov 24 09:29:23 crc kubenswrapper[4799]: I1124 09:29:23.931156 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-utilities" (OuterVolumeSpecName: "utilities") pod "c82c0477-5daa-4cce-a62d-05bb4f64f49b" (UID: "c82c0477-5daa-4cce-a62d-05bb4f64f49b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:29:23 crc kubenswrapper[4799]: I1124 09:29:23.947311 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c82c0477-5daa-4cce-a62d-05bb4f64f49b-kube-api-access-x6vs4" (OuterVolumeSpecName: "kube-api-access-x6vs4") pod "c82c0477-5daa-4cce-a62d-05bb4f64f49b" (UID: "c82c0477-5daa-4cce-a62d-05bb4f64f49b"). InnerVolumeSpecName "kube-api-access-x6vs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.017669 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c82c0477-5daa-4cce-a62d-05bb4f64f49b" (UID: "c82c0477-5daa-4cce-a62d-05bb4f64f49b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.031065 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6vs4\" (UniqueName: \"kubernetes.io/projected/c82c0477-5daa-4cce-a62d-05bb4f64f49b-kube-api-access-x6vs4\") on node \"crc\" DevicePath \"\"" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.031095 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.031109 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c82c0477-5daa-4cce-a62d-05bb4f64f49b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.302347 4799 generic.go:334] "Generic (PLEG): container finished" podID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerID="48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964" exitCode=0 Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.302464 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xs5gp" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.302405 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs5gp" event={"ID":"c82c0477-5daa-4cce-a62d-05bb4f64f49b","Type":"ContainerDied","Data":"48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964"} Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.302919 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs5gp" event={"ID":"c82c0477-5daa-4cce-a62d-05bb4f64f49b","Type":"ContainerDied","Data":"af55782a8c0139af72633cb852be76e2825417a25238d017e403752cb3b1e7c7"} Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.302947 4799 scope.go:117] "RemoveContainer" containerID="48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.344239 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xs5gp"] Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.344452 4799 scope.go:117] "RemoveContainer" containerID="9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.357943 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xs5gp"] Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.374150 4799 scope.go:117] "RemoveContainer" containerID="7abe6ba20d98418c262adcb19e266f84efe5b63249ff9f7fb9d0ae1b22e005e6" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.477042 4799 scope.go:117] "RemoveContainer" containerID="48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964" Nov 24 09:29:24 crc kubenswrapper[4799]: E1124 09:29:24.481014 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964\": container with ID starting with 48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964 not found: ID does not exist" containerID="48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.481073 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964"} err="failed to get container status \"48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964\": rpc error: code = NotFound desc = could not find container \"48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964\": container with ID starting with 48f7198a328208a1f22acc9fc79900e1ec3c68ac1425c9c12c0561a1b6192964 not found: ID does not exist" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.481104 4799 scope.go:117] "RemoveContainer" containerID="9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387" Nov 24 09:29:24 crc kubenswrapper[4799]: E1124 09:29:24.498053 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387\": container with ID starting with 9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387 not found: ID does not exist" containerID="9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.498107 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387"} err="failed to get container status \"9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387\": rpc error: code = NotFound desc = could not find container \"9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387\": container with ID starting with 9146030c3eaf32f98c20060bafa9e9e6c7866e2127a6c005c3192d2883fcf387 not found: ID does not exist" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.498140 4799 scope.go:117] "RemoveContainer" containerID="7abe6ba20d98418c262adcb19e266f84efe5b63249ff9f7fb9d0ae1b22e005e6" Nov 24 09:29:24 crc kubenswrapper[4799]: E1124 09:29:24.506024 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7abe6ba20d98418c262adcb19e266f84efe5b63249ff9f7fb9d0ae1b22e005e6\": container with ID starting with 7abe6ba20d98418c262adcb19e266f84efe5b63249ff9f7fb9d0ae1b22e005e6 not found: ID does not exist" containerID="7abe6ba20d98418c262adcb19e266f84efe5b63249ff9f7fb9d0ae1b22e005e6" Nov 24 09:29:24 crc kubenswrapper[4799]: I1124 09:29:24.506077 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7abe6ba20d98418c262adcb19e266f84efe5b63249ff9f7fb9d0ae1b22e005e6"} err="failed to get container status \"7abe6ba20d98418c262adcb19e266f84efe5b63249ff9f7fb9d0ae1b22e005e6\": rpc error: code = NotFound desc = could not find container \"7abe6ba20d98418c262adcb19e266f84efe5b63249ff9f7fb9d0ae1b22e005e6\": container with ID starting with 7abe6ba20d98418c262adcb19e266f84efe5b63249ff9f7fb9d0ae1b22e005e6 not found: ID does not exist" Nov 24 09:29:25 crc kubenswrapper[4799]: I1124 09:29:25.650708 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" path="/var/lib/kubelet/pods/c82c0477-5daa-4cce-a62d-05bb4f64f49b/volumes" Nov 24 09:29:50 crc kubenswrapper[4799]: I1124 09:29:50.401063 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:29:50 crc kubenswrapper[4799]: I1124 09:29:50.401615 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.166397 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj"] Nov 24 09:30:00 crc kubenswrapper[4799]: E1124 09:30:00.168398 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerName="extract-content" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.168423 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerName="extract-content" Nov 24 09:30:00 crc kubenswrapper[4799]: E1124 09:30:00.168458 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerName="extract-utilities" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.168467 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerName="extract-utilities" Nov 24 09:30:00 crc kubenswrapper[4799]: E1124 09:30:00.168488 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerName="registry-server" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.168495 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerName="registry-server" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.168761 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c82c0477-5daa-4cce-a62d-05bb4f64f49b" containerName="registry-server" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.169770 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.172429 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.172429 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.182949 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj"] Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.194203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a312fbf-3397-4f33-9a18-8d3810966f42-secret-volume\") pod \"collect-profiles-29399610-sjbmj\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.194259 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf9l9\" (UniqueName: \"kubernetes.io/projected/6a312fbf-3397-4f33-9a18-8d3810966f42-kube-api-access-sf9l9\") pod \"collect-profiles-29399610-sjbmj\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.194409 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a312fbf-3397-4f33-9a18-8d3810966f42-config-volume\") pod \"collect-profiles-29399610-sjbmj\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.296962 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a312fbf-3397-4f33-9a18-8d3810966f42-config-volume\") pod \"collect-profiles-29399610-sjbmj\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.297157 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a312fbf-3397-4f33-9a18-8d3810966f42-secret-volume\") pod \"collect-profiles-29399610-sjbmj\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.297184 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf9l9\" (UniqueName: \"kubernetes.io/projected/6a312fbf-3397-4f33-9a18-8d3810966f42-kube-api-access-sf9l9\") pod \"collect-profiles-29399610-sjbmj\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.298644 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a312fbf-3397-4f33-9a18-8d3810966f42-config-volume\") pod \"collect-profiles-29399610-sjbmj\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.305500 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a312fbf-3397-4f33-9a18-8d3810966f42-secret-volume\") pod \"collect-profiles-29399610-sjbmj\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.318805 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf9l9\" (UniqueName: \"kubernetes.io/projected/6a312fbf-3397-4f33-9a18-8d3810966f42-kube-api-access-sf9l9\") pod \"collect-profiles-29399610-sjbmj\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.493827 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:00 crc kubenswrapper[4799]: I1124 09:30:00.994091 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj"] Nov 24 09:30:01 crc kubenswrapper[4799]: I1124 09:30:01.681442 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" event={"ID":"6a312fbf-3397-4f33-9a18-8d3810966f42","Type":"ContainerStarted","Data":"8a4bd04a48080432ba3e253fe5ff5e39998c4e10c5a52b67673ca9cf9f0e6edb"} Nov 24 09:30:01 crc kubenswrapper[4799]: I1124 09:30:01.681718 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" event={"ID":"6a312fbf-3397-4f33-9a18-8d3810966f42","Type":"ContainerStarted","Data":"4376c74156c3a30149de5468a6f2fb0a1cb6b82cd1287b9b992e1fca9b253b32"} Nov 24 09:30:02 crc kubenswrapper[4799]: I1124 09:30:02.690979 4799 generic.go:334] "Generic (PLEG): container finished" podID="6a312fbf-3397-4f33-9a18-8d3810966f42" containerID="8a4bd04a48080432ba3e253fe5ff5e39998c4e10c5a52b67673ca9cf9f0e6edb" exitCode=0 Nov 24 09:30:02 crc kubenswrapper[4799]: I1124 09:30:02.691276 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" event={"ID":"6a312fbf-3397-4f33-9a18-8d3810966f42","Type":"ContainerDied","Data":"8a4bd04a48080432ba3e253fe5ff5e39998c4e10c5a52b67673ca9cf9f0e6edb"} Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.151895 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.187383 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf9l9\" (UniqueName: \"kubernetes.io/projected/6a312fbf-3397-4f33-9a18-8d3810966f42-kube-api-access-sf9l9\") pod \"6a312fbf-3397-4f33-9a18-8d3810966f42\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.187537 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a312fbf-3397-4f33-9a18-8d3810966f42-config-volume\") pod \"6a312fbf-3397-4f33-9a18-8d3810966f42\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.187661 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a312fbf-3397-4f33-9a18-8d3810966f42-secret-volume\") pod \"6a312fbf-3397-4f33-9a18-8d3810966f42\" (UID: \"6a312fbf-3397-4f33-9a18-8d3810966f42\") " Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.188612 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a312fbf-3397-4f33-9a18-8d3810966f42-config-volume" (OuterVolumeSpecName: "config-volume") pod "6a312fbf-3397-4f33-9a18-8d3810966f42" (UID: "6a312fbf-3397-4f33-9a18-8d3810966f42"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.193893 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a312fbf-3397-4f33-9a18-8d3810966f42-kube-api-access-sf9l9" (OuterVolumeSpecName: "kube-api-access-sf9l9") pod "6a312fbf-3397-4f33-9a18-8d3810966f42" (UID: "6a312fbf-3397-4f33-9a18-8d3810966f42"). InnerVolumeSpecName "kube-api-access-sf9l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.194002 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a312fbf-3397-4f33-9a18-8d3810966f42-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6a312fbf-3397-4f33-9a18-8d3810966f42" (UID: "6a312fbf-3397-4f33-9a18-8d3810966f42"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.293994 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a312fbf-3397-4f33-9a18-8d3810966f42-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.294027 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a312fbf-3397-4f33-9a18-8d3810966f42-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.294036 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf9l9\" (UniqueName: \"kubernetes.io/projected/6a312fbf-3397-4f33-9a18-8d3810966f42-kube-api-access-sf9l9\") on node \"crc\" DevicePath \"\"" Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.717140 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" event={"ID":"6a312fbf-3397-4f33-9a18-8d3810966f42","Type":"ContainerDied","Data":"4376c74156c3a30149de5468a6f2fb0a1cb6b82cd1287b9b992e1fca9b253b32"} Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.717189 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-sjbmj" Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.717196 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4376c74156c3a30149de5468a6f2fb0a1cb6b82cd1287b9b992e1fca9b253b32" Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.779960 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7"] Nov 24 09:30:04 crc kubenswrapper[4799]: I1124 09:30:04.791653 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-bs4v7"] Nov 24 09:30:05 crc kubenswrapper[4799]: I1124 09:30:05.642088 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5971752e-f0b4-4ece-b52f-e8a004ea0ff6" path="/var/lib/kubelet/pods/5971752e-f0b4-4ece-b52f-e8a004ea0ff6/volumes" Nov 24 09:30:13 crc kubenswrapper[4799]: I1124 09:30:13.011355 4799 scope.go:117] "RemoveContainer" containerID="c19b58ad3cf69e86484938a1bcbb6e461d41cf31fef782a4dfa0dbdd0af08dce" Nov 24 09:30:19 crc kubenswrapper[4799]: I1124 09:30:19.858863 4799 generic.go:334] "Generic (PLEG): container finished" podID="8bbbf471-6b23-485e-9118-881aec43406c" containerID="1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5" exitCode=0 Nov 24 09:30:19 crc kubenswrapper[4799]: I1124 09:30:19.858952 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mwth7/must-gather-gs8f2" event={"ID":"8bbbf471-6b23-485e-9118-881aec43406c","Type":"ContainerDied","Data":"1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5"} Nov 24 09:30:19 crc kubenswrapper[4799]: I1124 09:30:19.860396 4799 scope.go:117] "RemoveContainer" containerID="1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5" Nov 24 09:30:20 crc kubenswrapper[4799]: I1124 09:30:20.175737 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mwth7_must-gather-gs8f2_8bbbf471-6b23-485e-9118-881aec43406c/gather/0.log" Nov 24 09:30:20 crc kubenswrapper[4799]: I1124 09:30:20.400946 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:30:20 crc kubenswrapper[4799]: I1124 09:30:20.401033 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.198144 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mwth7/must-gather-gs8f2"] Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.198967 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mwth7/must-gather-gs8f2" podUID="8bbbf471-6b23-485e-9118-881aec43406c" containerName="copy" containerID="cri-o://8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b" gracePeriod=2 Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.214935 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mwth7/must-gather-gs8f2"] Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.673551 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mwth7_must-gather-gs8f2_8bbbf471-6b23-485e-9118-881aec43406c/copy/0.log" Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.674526 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/must-gather-gs8f2" Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.799561 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8bbbf471-6b23-485e-9118-881aec43406c-must-gather-output\") pod \"8bbbf471-6b23-485e-9118-881aec43406c\" (UID: \"8bbbf471-6b23-485e-9118-881aec43406c\") " Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.799869 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-962c4\" (UniqueName: \"kubernetes.io/projected/8bbbf471-6b23-485e-9118-881aec43406c-kube-api-access-962c4\") pod \"8bbbf471-6b23-485e-9118-881aec43406c\" (UID: \"8bbbf471-6b23-485e-9118-881aec43406c\") " Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.807901 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bbbf471-6b23-485e-9118-881aec43406c-kube-api-access-962c4" (OuterVolumeSpecName: "kube-api-access-962c4") pod "8bbbf471-6b23-485e-9118-881aec43406c" (UID: "8bbbf471-6b23-485e-9118-881aec43406c"). InnerVolumeSpecName "kube-api-access-962c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.902591 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-962c4\" (UniqueName: \"kubernetes.io/projected/8bbbf471-6b23-485e-9118-881aec43406c-kube-api-access-962c4\") on node \"crc\" DevicePath \"\"" Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.958677 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mwth7_must-gather-gs8f2_8bbbf471-6b23-485e-9118-881aec43406c/copy/0.log" Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.959004 4799 generic.go:334] "Generic (PLEG): container finished" podID="8bbbf471-6b23-485e-9118-881aec43406c" containerID="8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b" exitCode=143 Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.959061 4799 scope.go:117] "RemoveContainer" containerID="8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b" Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.959086 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mwth7/must-gather-gs8f2" Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.974032 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bbbf471-6b23-485e-9118-881aec43406c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "8bbbf471-6b23-485e-9118-881aec43406c" (UID: "8bbbf471-6b23-485e-9118-881aec43406c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:30:28 crc kubenswrapper[4799]: I1124 09:30:28.982351 4799 scope.go:117] "RemoveContainer" containerID="1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5" Nov 24 09:30:29 crc kubenswrapper[4799]: I1124 09:30:29.004659 4799 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8bbbf471-6b23-485e-9118-881aec43406c-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 09:30:29 crc kubenswrapper[4799]: I1124 09:30:29.114814 4799 scope.go:117] "RemoveContainer" containerID="8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b" Nov 24 09:30:29 crc kubenswrapper[4799]: E1124 09:30:29.116002 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b\": container with ID starting with 8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b not found: ID does not exist" containerID="8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b" Nov 24 09:30:29 crc kubenswrapper[4799]: I1124 09:30:29.116054 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b"} err="failed to get container status \"8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b\": rpc error: code = NotFound desc = could not find container \"8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b\": container with ID starting with 8ffb470354d0b42f761250d91c0ac92f50e5b494b3feb5f2f13baeffe9035c6b not found: ID does not exist" Nov 24 09:30:29 crc kubenswrapper[4799]: I1124 09:30:29.116089 4799 scope.go:117] "RemoveContainer" containerID="1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5" Nov 24 09:30:29 crc kubenswrapper[4799]: E1124 09:30:29.116427 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5\": container with ID starting with 1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5 not found: ID does not exist" containerID="1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5" Nov 24 09:30:29 crc kubenswrapper[4799]: I1124 09:30:29.116463 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5"} err="failed to get container status \"1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5\": rpc error: code = NotFound desc = could not find container \"1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5\": container with ID starting with 1e1f6cb9b7c73faafd7351fc533cb7680891882c78ac7643c2570eecfe5b82e5 not found: ID does not exist" Nov 24 09:30:29 crc kubenswrapper[4799]: I1124 09:30:29.642206 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bbbf471-6b23-485e-9118-881aec43406c" path="/var/lib/kubelet/pods/8bbbf471-6b23-485e-9118-881aec43406c/volumes" Nov 24 09:30:50 crc kubenswrapper[4799]: I1124 09:30:50.401037 4799 patch_prober.go:28] interesting pod/machine-config-daemon-rzk7n container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:30:50 crc kubenswrapper[4799]: I1124 09:30:50.401526 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:30:50 crc kubenswrapper[4799]: I1124 09:30:50.401572 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" Nov 24 09:30:50 crc kubenswrapper[4799]: I1124 09:30:50.402379 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532"} pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:30:50 crc kubenswrapper[4799]: I1124 09:30:50.402441 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" containerName="machine-config-daemon" containerID="cri-o://626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" gracePeriod=600 Nov 24 09:30:50 crc kubenswrapper[4799]: E1124 09:30:50.536717 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:30:51 crc kubenswrapper[4799]: I1124 09:30:51.175401 4799 generic.go:334] "Generic (PLEG): container finished" podID="354ebd50-7a67-481d-86db-4b3bf4753161" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" exitCode=0 Nov 24 09:30:51 crc kubenswrapper[4799]: I1124 09:30:51.175491 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" event={"ID":"354ebd50-7a67-481d-86db-4b3bf4753161","Type":"ContainerDied","Data":"626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532"} Nov 24 09:30:51 crc kubenswrapper[4799]: I1124 09:30:51.175741 4799 scope.go:117] "RemoveContainer" containerID="517c10c0986c1088f2ae888718c16d2c685822643c593b40da162bf2f3673af3" Nov 24 09:30:51 crc kubenswrapper[4799]: I1124 09:30:51.176491 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:30:51 crc kubenswrapper[4799]: E1124 09:30:51.176813 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:31:03 crc kubenswrapper[4799]: I1124 09:31:03.628265 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:31:03 crc kubenswrapper[4799]: E1124 09:31:03.629123 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:31:15 crc kubenswrapper[4799]: I1124 09:31:15.637419 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:31:15 crc kubenswrapper[4799]: E1124 09:31:15.638606 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:31:26 crc kubenswrapper[4799]: I1124 09:31:26.629671 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:31:26 crc kubenswrapper[4799]: E1124 09:31:26.632776 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:31:39 crc kubenswrapper[4799]: I1124 09:31:39.628588 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:31:39 crc kubenswrapper[4799]: E1124 09:31:39.629362 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.233673 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lb58l"] Nov 24 09:31:46 crc kubenswrapper[4799]: E1124 09:31:46.234414 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a312fbf-3397-4f33-9a18-8d3810966f42" containerName="collect-profiles" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.234428 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a312fbf-3397-4f33-9a18-8d3810966f42" containerName="collect-profiles" Nov 24 09:31:46 crc kubenswrapper[4799]: E1124 09:31:46.234455 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bbbf471-6b23-485e-9118-881aec43406c" containerName="gather" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.234461 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bbbf471-6b23-485e-9118-881aec43406c" containerName="gather" Nov 24 09:31:46 crc kubenswrapper[4799]: E1124 09:31:46.234467 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bbbf471-6b23-485e-9118-881aec43406c" containerName="copy" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.234473 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bbbf471-6b23-485e-9118-881aec43406c" containerName="copy" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.234678 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bbbf471-6b23-485e-9118-881aec43406c" containerName="gather" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.234697 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bbbf471-6b23-485e-9118-881aec43406c" containerName="copy" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.234710 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a312fbf-3397-4f33-9a18-8d3810966f42" containerName="collect-profiles" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.236479 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.253246 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-utilities\") pod \"certified-operators-lb58l\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.253451 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-catalog-content\") pod \"certified-operators-lb58l\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.254488 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxl9h\" (UniqueName: \"kubernetes.io/projected/2b997468-7605-4b9d-adee-56a9a3b6d98d-kube-api-access-pxl9h\") pod \"certified-operators-lb58l\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.262441 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lb58l"] Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.356838 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxl9h\" (UniqueName: \"kubernetes.io/projected/2b997468-7605-4b9d-adee-56a9a3b6d98d-kube-api-access-pxl9h\") pod \"certified-operators-lb58l\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.357055 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-utilities\") pod \"certified-operators-lb58l\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.357188 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-catalog-content\") pod \"certified-operators-lb58l\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.357698 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-utilities\") pod \"certified-operators-lb58l\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.357737 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-catalog-content\") pod \"certified-operators-lb58l\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.378701 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxl9h\" (UniqueName: \"kubernetes.io/projected/2b997468-7605-4b9d-adee-56a9a3b6d98d-kube-api-access-pxl9h\") pod \"certified-operators-lb58l\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:46 crc kubenswrapper[4799]: I1124 09:31:46.558746 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:47 crc kubenswrapper[4799]: I1124 09:31:47.060797 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lb58l"] Nov 24 09:31:47 crc kubenswrapper[4799]: I1124 09:31:47.768874 4799 generic.go:334] "Generic (PLEG): container finished" podID="2b997468-7605-4b9d-adee-56a9a3b6d98d" containerID="265746ab0b3c0ba395855eb65c8f4b16d2dcd4dfe3de7d9160af233b3d6754bf" exitCode=0 Nov 24 09:31:47 crc kubenswrapper[4799]: I1124 09:31:47.768971 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lb58l" event={"ID":"2b997468-7605-4b9d-adee-56a9a3b6d98d","Type":"ContainerDied","Data":"265746ab0b3c0ba395855eb65c8f4b16d2dcd4dfe3de7d9160af233b3d6754bf"} Nov 24 09:31:47 crc kubenswrapper[4799]: I1124 09:31:47.769282 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lb58l" event={"ID":"2b997468-7605-4b9d-adee-56a9a3b6d98d","Type":"ContainerStarted","Data":"d133dd470a9a435f916b4e076e9ad1cfb95a34453fb672c6ea4e5be2e55425a3"} Nov 24 09:31:47 crc kubenswrapper[4799]: I1124 09:31:47.771028 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:31:49 crc kubenswrapper[4799]: I1124 09:31:49.799283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lb58l" event={"ID":"2b997468-7605-4b9d-adee-56a9a3b6d98d","Type":"ContainerStarted","Data":"1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b"} Nov 24 09:31:50 crc kubenswrapper[4799]: I1124 09:31:50.810543 4799 generic.go:334] "Generic (PLEG): container finished" podID="2b997468-7605-4b9d-adee-56a9a3b6d98d" containerID="1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b" exitCode=0 Nov 24 09:31:50 crc kubenswrapper[4799]: I1124 09:31:50.810601 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lb58l" event={"ID":"2b997468-7605-4b9d-adee-56a9a3b6d98d","Type":"ContainerDied","Data":"1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b"} Nov 24 09:31:51 crc kubenswrapper[4799]: I1124 09:31:51.822688 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lb58l" event={"ID":"2b997468-7605-4b9d-adee-56a9a3b6d98d","Type":"ContainerStarted","Data":"653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821"} Nov 24 09:31:51 crc kubenswrapper[4799]: I1124 09:31:51.853520 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lb58l" podStartSLOduration=2.418136948 podStartE2EDuration="5.853499082s" podCreationTimestamp="2025-11-24 09:31:46 +0000 UTC" firstStartedPulling="2025-11-24 09:31:47.770750647 +0000 UTC m=+9853.426733111" lastFinishedPulling="2025-11-24 09:31:51.206112771 +0000 UTC m=+9856.862095245" observedRunningTime="2025-11-24 09:31:51.847713467 +0000 UTC m=+9857.503695941" watchObservedRunningTime="2025-11-24 09:31:51.853499082 +0000 UTC m=+9857.509481566" Nov 24 09:31:53 crc kubenswrapper[4799]: I1124 09:31:53.629211 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:31:53 crc kubenswrapper[4799]: E1124 09:31:53.630135 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:31:56 crc kubenswrapper[4799]: I1124 09:31:56.559055 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:56 crc kubenswrapper[4799]: I1124 09:31:56.560162 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:56 crc kubenswrapper[4799]: I1124 09:31:56.622114 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:56 crc kubenswrapper[4799]: I1124 09:31:56.918702 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:31:59 crc kubenswrapper[4799]: I1124 09:31:59.017825 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lb58l"] Nov 24 09:31:59 crc kubenswrapper[4799]: I1124 09:31:59.896161 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lb58l" podUID="2b997468-7605-4b9d-adee-56a9a3b6d98d" containerName="registry-server" containerID="cri-o://653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821" gracePeriod=2 Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.402962 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.471803 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-utilities\") pod \"2b997468-7605-4b9d-adee-56a9a3b6d98d\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.471980 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-catalog-content\") pod \"2b997468-7605-4b9d-adee-56a9a3b6d98d\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.472032 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxl9h\" (UniqueName: \"kubernetes.io/projected/2b997468-7605-4b9d-adee-56a9a3b6d98d-kube-api-access-pxl9h\") pod \"2b997468-7605-4b9d-adee-56a9a3b6d98d\" (UID: \"2b997468-7605-4b9d-adee-56a9a3b6d98d\") " Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.472978 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-utilities" (OuterVolumeSpecName: "utilities") pod "2b997468-7605-4b9d-adee-56a9a3b6d98d" (UID: "2b997468-7605-4b9d-adee-56a9a3b6d98d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.477281 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b997468-7605-4b9d-adee-56a9a3b6d98d-kube-api-access-pxl9h" (OuterVolumeSpecName: "kube-api-access-pxl9h") pod "2b997468-7605-4b9d-adee-56a9a3b6d98d" (UID: "2b997468-7605-4b9d-adee-56a9a3b6d98d"). InnerVolumeSpecName "kube-api-access-pxl9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.523512 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b997468-7605-4b9d-adee-56a9a3b6d98d" (UID: "2b997468-7605-4b9d-adee-56a9a3b6d98d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.574098 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxl9h\" (UniqueName: \"kubernetes.io/projected/2b997468-7605-4b9d-adee-56a9a3b6d98d-kube-api-access-pxl9h\") on node \"crc\" DevicePath \"\"" Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.574135 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.574146 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b997468-7605-4b9d-adee-56a9a3b6d98d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.906750 4799 generic.go:334] "Generic (PLEG): container finished" podID="2b997468-7605-4b9d-adee-56a9a3b6d98d" containerID="653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821" exitCode=0 Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.906793 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lb58l" event={"ID":"2b997468-7605-4b9d-adee-56a9a3b6d98d","Type":"ContainerDied","Data":"653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821"} Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.906821 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lb58l" event={"ID":"2b997468-7605-4b9d-adee-56a9a3b6d98d","Type":"ContainerDied","Data":"d133dd470a9a435f916b4e076e9ad1cfb95a34453fb672c6ea4e5be2e55425a3"} Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.906838 4799 scope.go:117] "RemoveContainer" containerID="653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821" Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.907013 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lb58l" Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.933899 4799 scope.go:117] "RemoveContainer" containerID="1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b" Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.945456 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lb58l"] Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.954106 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lb58l"] Nov 24 09:32:00 crc kubenswrapper[4799]: I1124 09:32:00.973703 4799 scope.go:117] "RemoveContainer" containerID="265746ab0b3c0ba395855eb65c8f4b16d2dcd4dfe3de7d9160af233b3d6754bf" Nov 24 09:32:01 crc kubenswrapper[4799]: I1124 09:32:01.004574 4799 scope.go:117] "RemoveContainer" containerID="653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821" Nov 24 09:32:01 crc kubenswrapper[4799]: E1124 09:32:01.005147 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821\": container with ID starting with 653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821 not found: ID does not exist" containerID="653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821" Nov 24 09:32:01 crc kubenswrapper[4799]: I1124 09:32:01.005185 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821"} err="failed to get container status \"653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821\": rpc error: code = NotFound desc = could not find container \"653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821\": container with ID starting with 653a15e07d3ed4b0381f2b254be7ba8b98e5d158d20b236b5515604cd73fb821 not found: ID does not exist" Nov 24 09:32:01 crc kubenswrapper[4799]: I1124 09:32:01.005215 4799 scope.go:117] "RemoveContainer" containerID="1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b" Nov 24 09:32:01 crc kubenswrapper[4799]: E1124 09:32:01.005691 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b\": container with ID starting with 1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b not found: ID does not exist" containerID="1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b" Nov 24 09:32:01 crc kubenswrapper[4799]: I1124 09:32:01.005720 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b"} err="failed to get container status \"1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b\": rpc error: code = NotFound desc = could not find container \"1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b\": container with ID starting with 1f872aefe091201a963f0b838f102c6c3867050698bbcf04049992f2348ca30b not found: ID does not exist" Nov 24 09:32:01 crc kubenswrapper[4799]: I1124 09:32:01.005738 4799 scope.go:117] "RemoveContainer" containerID="265746ab0b3c0ba395855eb65c8f4b16d2dcd4dfe3de7d9160af233b3d6754bf" Nov 24 09:32:01 crc kubenswrapper[4799]: E1124 09:32:01.006124 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"265746ab0b3c0ba395855eb65c8f4b16d2dcd4dfe3de7d9160af233b3d6754bf\": container with ID starting with 265746ab0b3c0ba395855eb65c8f4b16d2dcd4dfe3de7d9160af233b3d6754bf not found: ID does not exist" containerID="265746ab0b3c0ba395855eb65c8f4b16d2dcd4dfe3de7d9160af233b3d6754bf" Nov 24 09:32:01 crc kubenswrapper[4799]: I1124 09:32:01.006149 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"265746ab0b3c0ba395855eb65c8f4b16d2dcd4dfe3de7d9160af233b3d6754bf"} err="failed to get container status \"265746ab0b3c0ba395855eb65c8f4b16d2dcd4dfe3de7d9160af233b3d6754bf\": rpc error: code = NotFound desc = could not find container \"265746ab0b3c0ba395855eb65c8f4b16d2dcd4dfe3de7d9160af233b3d6754bf\": container with ID starting with 265746ab0b3c0ba395855eb65c8f4b16d2dcd4dfe3de7d9160af233b3d6754bf not found: ID does not exist" Nov 24 09:32:01 crc kubenswrapper[4799]: I1124 09:32:01.639002 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b997468-7605-4b9d-adee-56a9a3b6d98d" path="/var/lib/kubelet/pods/2b997468-7605-4b9d-adee-56a9a3b6d98d/volumes" Nov 24 09:32:07 crc kubenswrapper[4799]: I1124 09:32:07.628779 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:32:07 crc kubenswrapper[4799]: E1124 09:32:07.630088 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:32:18 crc kubenswrapper[4799]: I1124 09:32:18.628981 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:32:18 crc kubenswrapper[4799]: E1124 09:32:18.629742 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:32:32 crc kubenswrapper[4799]: I1124 09:32:32.628320 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:32:32 crc kubenswrapper[4799]: E1124 09:32:32.629183 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:32:47 crc kubenswrapper[4799]: I1124 09:32:47.632705 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:32:47 crc kubenswrapper[4799]: E1124 09:32:47.633592 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:33:02 crc kubenswrapper[4799]: I1124 09:33:02.632942 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:33:02 crc kubenswrapper[4799]: E1124 09:33:02.634627 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:33:17 crc kubenswrapper[4799]: I1124 09:33:17.628912 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:33:17 crc kubenswrapper[4799]: E1124 09:33:17.629711 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:33:31 crc kubenswrapper[4799]: I1124 09:33:31.627880 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:33:31 crc kubenswrapper[4799]: E1124 09:33:31.628690 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" Nov 24 09:33:44 crc kubenswrapper[4799]: I1124 09:33:44.628022 4799 scope.go:117] "RemoveContainer" containerID="626f611374b5d4b874e31c1bb7c504f8fd7d79465e7b2a9c2d5d2e3d2324b532" Nov 24 09:33:44 crc kubenswrapper[4799]: E1124 09:33:44.628805 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rzk7n_openshift-machine-config-operator(354ebd50-7a67-481d-86db-4b3bf4753161)\"" pod="openshift-machine-config-operator/machine-config-daemon-rzk7n" podUID="354ebd50-7a67-481d-86db-4b3bf4753161" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111023213024432 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111023214017350 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015110777255016517 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015110777255015467 5ustar corecore